Aug 13 07:20:46.018996 kernel: Linux version 6.6.100-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Aug 12 22:14:58 -00 2025 Aug 13 07:20:46.019011 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:20:46.019019 kernel: BIOS-provided physical RAM map: Aug 13 07:20:46.019023 kernel: BIOS-e820: [mem 0x0000000000000000-0x00000000000997ff] usable Aug 13 07:20:46.019027 kernel: BIOS-e820: [mem 0x0000000000099800-0x000000000009ffff] reserved Aug 13 07:20:46.019031 kernel: BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved Aug 13 07:20:46.019035 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003fffffff] usable Aug 13 07:20:46.019040 kernel: BIOS-e820: [mem 0x0000000040000000-0x00000000403fffff] reserved Aug 13 07:20:46.019044 kernel: BIOS-e820: [mem 0x0000000040400000-0x00000000819c6fff] usable Aug 13 07:20:46.019048 kernel: BIOS-e820: [mem 0x00000000819c7000-0x00000000819c7fff] ACPI NVS Aug 13 07:20:46.019052 kernel: BIOS-e820: [mem 0x00000000819c8000-0x00000000819c8fff] reserved Aug 13 07:20:46.019057 kernel: BIOS-e820: [mem 0x00000000819c9000-0x000000008afcdfff] usable Aug 13 07:20:46.019061 kernel: BIOS-e820: [mem 0x000000008afce000-0x000000008c0b2fff] reserved Aug 13 07:20:46.019068 kernel: BIOS-e820: [mem 0x000000008c0b3000-0x000000008c23bfff] usable Aug 13 07:20:46.019073 kernel: BIOS-e820: [mem 0x000000008c23c000-0x000000008c66dfff] ACPI NVS Aug 13 07:20:46.019078 kernel: BIOS-e820: [mem 0x000000008c66e000-0x000000008eefefff] reserved Aug 13 07:20:46.019102 kernel: BIOS-e820: [mem 0x000000008eeff000-0x000000008eefffff] usable Aug 13 07:20:46.019107 kernel: BIOS-e820: [mem 0x000000008ef00000-0x000000008fffffff] reserved Aug 13 07:20:46.019112 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Aug 13 07:20:46.019117 kernel: BIOS-e820: [mem 0x00000000fe000000-0x00000000fe010fff] reserved Aug 13 07:20:46.019135 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec00fff] reserved Aug 13 07:20:46.019139 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Aug 13 07:20:46.019144 kernel: BIOS-e820: [mem 0x00000000ff000000-0x00000000ffffffff] reserved Aug 13 07:20:46.019148 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000086effffff] usable Aug 13 07:20:46.019153 kernel: NX (Execute Disable) protection: active Aug 13 07:20:46.019158 kernel: APIC: Static calls initialized Aug 13 07:20:46.019162 kernel: SMBIOS 3.2.1 present. Aug 13 07:20:46.019167 kernel: DMI: Supermicro SYS-5019C-MR-PH004/X11SCM-F, BIOS 2.6 12/03/2024 Aug 13 07:20:46.019173 kernel: tsc: Detected 3400.000 MHz processor Aug 13 07:20:46.019177 kernel: tsc: Detected 3399.906 MHz TSC Aug 13 07:20:46.019182 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Aug 13 07:20:46.019187 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Aug 13 07:20:46.019192 kernel: last_pfn = 0x86f000 max_arch_pfn = 0x400000000 Aug 13 07:20:46.019197 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 23), built from 10 variable MTRRs Aug 13 07:20:46.019201 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Aug 13 07:20:46.019206 kernel: last_pfn = 0x8ef00 max_arch_pfn = 0x400000000 Aug 13 07:20:46.019211 kernel: Using GB pages for direct mapping Aug 13 07:20:46.019217 kernel: ACPI: Early table checksum verification disabled Aug 13 07:20:46.019221 kernel: ACPI: RSDP 0x00000000000F05B0 000024 (v02 SUPERM) Aug 13 07:20:46.019226 kernel: ACPI: XSDT 0x000000008C54F0C8 00010C (v01 SUPERM SUPERM 01072009 AMI 00010013) Aug 13 07:20:46.019233 kernel: ACPI: FACP 0x000000008C58B670 000114 (v06 01072009 AMI 00010013) Aug 13 07:20:46.019238 kernel: ACPI: DSDT 0x000000008C54F268 03C404 (v02 SUPERM SMCI--MB 01072009 INTL 20160527) Aug 13 07:20:46.019243 kernel: ACPI: FACS 0x000000008C66DF80 000040 Aug 13 07:20:46.019248 kernel: ACPI: APIC 0x000000008C58B788 00012C (v04 01072009 AMI 00010013) Aug 13 07:20:46.019254 kernel: ACPI: FPDT 0x000000008C58B8B8 000044 (v01 01072009 AMI 00010013) Aug 13 07:20:46.019259 kernel: ACPI: FIDT 0x000000008C58B900 00009C (v01 SUPERM SMCI--MB 01072009 AMI 00010013) Aug 13 07:20:46.019264 kernel: ACPI: MCFG 0x000000008C58B9A0 00003C (v01 SUPERM SMCI--MB 01072009 MSFT 00000097) Aug 13 07:20:46.019269 kernel: ACPI: SPMI 0x000000008C58B9E0 000041 (v05 SUPERM SMCI--MB 00000000 AMI. 00000000) Aug 13 07:20:46.019274 kernel: ACPI: SSDT 0x000000008C58BA28 001B1C (v02 CpuRef CpuSsdt 00003000 INTL 20160527) Aug 13 07:20:46.019279 kernel: ACPI: SSDT 0x000000008C58D548 0031C6 (v02 SaSsdt SaSsdt 00003000 INTL 20160527) Aug 13 07:20:46.019283 kernel: ACPI: SSDT 0x000000008C590710 00232B (v02 PegSsd PegSsdt 00001000 INTL 20160527) Aug 13 07:20:46.019289 kernel: ACPI: HPET 0x000000008C592A40 000038 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019294 kernel: ACPI: SSDT 0x000000008C592A78 000FAE (v02 SUPERM Ther_Rvp 00001000 INTL 20160527) Aug 13 07:20:46.019299 kernel: ACPI: SSDT 0x000000008C593A28 0008F4 (v02 INTEL xh_mossb 00000000 INTL 20160527) Aug 13 07:20:46.019304 kernel: ACPI: UEFI 0x000000008C594320 000042 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019309 kernel: ACPI: LPIT 0x000000008C594368 000094 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019314 kernel: ACPI: SSDT 0x000000008C594400 0027DE (v02 SUPERM PtidDevc 00001000 INTL 20160527) Aug 13 07:20:46.019319 kernel: ACPI: SSDT 0x000000008C596BE0 0014E2 (v02 SUPERM TbtTypeC 00000000 INTL 20160527) Aug 13 07:20:46.019324 kernel: ACPI: DBGP 0x000000008C5980C8 000034 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019329 kernel: ACPI: DBG2 0x000000008C598100 000054 (v00 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019335 kernel: ACPI: SSDT 0x000000008C598158 001B67 (v02 SUPERM UsbCTabl 00001000 INTL 20160527) Aug 13 07:20:46.019340 kernel: ACPI: DMAR 0x000000008C599CC0 000070 (v01 INTEL EDK2 00000002 01000013) Aug 13 07:20:46.019345 kernel: ACPI: SSDT 0x000000008C599D30 000144 (v02 Intel ADebTabl 00001000 INTL 20160527) Aug 13 07:20:46.019349 kernel: ACPI: TPM2 0x000000008C599E78 000034 (v04 SUPERM SMCI--MB 00000001 AMI 00000000) Aug 13 07:20:46.019354 kernel: ACPI: SSDT 0x000000008C599EB0 000D8F (v02 INTEL SpsNm 00000002 INTL 20160527) Aug 13 07:20:46.019359 kernel: ACPI: WSMT 0x000000008C59AC40 000028 (v01 SUPERM 01072009 AMI 00010013) Aug 13 07:20:46.019364 kernel: ACPI: EINJ 0x000000008C59AC68 000130 (v01 AMI AMI.EINJ 00000000 AMI. 00000000) Aug 13 07:20:46.019369 kernel: ACPI: ERST 0x000000008C59AD98 000230 (v01 AMIER AMI.ERST 00000000 AMI. 00000000) Aug 13 07:20:46.019375 kernel: ACPI: BERT 0x000000008C59AFC8 000030 (v01 AMI AMI.BERT 00000000 AMI. 00000000) Aug 13 07:20:46.019380 kernel: ACPI: HEST 0x000000008C59AFF8 00027C (v01 AMI AMI.HEST 00000000 AMI. 00000000) Aug 13 07:20:46.019385 kernel: ACPI: SSDT 0x000000008C59B278 000162 (v01 SUPERM SMCCDN 00000000 INTL 20181221) Aug 13 07:20:46.019390 kernel: ACPI: Reserving FACP table memory at [mem 0x8c58b670-0x8c58b783] Aug 13 07:20:46.019395 kernel: ACPI: Reserving DSDT table memory at [mem 0x8c54f268-0x8c58b66b] Aug 13 07:20:46.019400 kernel: ACPI: Reserving FACS table memory at [mem 0x8c66df80-0x8c66dfbf] Aug 13 07:20:46.019405 kernel: ACPI: Reserving APIC table memory at [mem 0x8c58b788-0x8c58b8b3] Aug 13 07:20:46.019410 kernel: ACPI: Reserving FPDT table memory at [mem 0x8c58b8b8-0x8c58b8fb] Aug 13 07:20:46.019415 kernel: ACPI: Reserving FIDT table memory at [mem 0x8c58b900-0x8c58b99b] Aug 13 07:20:46.019420 kernel: ACPI: Reserving MCFG table memory at [mem 0x8c58b9a0-0x8c58b9db] Aug 13 07:20:46.019425 kernel: ACPI: Reserving SPMI table memory at [mem 0x8c58b9e0-0x8c58ba20] Aug 13 07:20:46.019430 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58ba28-0x8c58d543] Aug 13 07:20:46.019435 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58d548-0x8c59070d] Aug 13 07:20:46.019440 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c590710-0x8c592a3a] Aug 13 07:20:46.019445 kernel: ACPI: Reserving HPET table memory at [mem 0x8c592a40-0x8c592a77] Aug 13 07:20:46.019450 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c592a78-0x8c593a25] Aug 13 07:20:46.019455 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c593a28-0x8c59431b] Aug 13 07:20:46.019459 kernel: ACPI: Reserving UEFI table memory at [mem 0x8c594320-0x8c594361] Aug 13 07:20:46.019465 kernel: ACPI: Reserving LPIT table memory at [mem 0x8c594368-0x8c5943fb] Aug 13 07:20:46.019470 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c594400-0x8c596bdd] Aug 13 07:20:46.019475 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c596be0-0x8c5980c1] Aug 13 07:20:46.019480 kernel: ACPI: Reserving DBGP table memory at [mem 0x8c5980c8-0x8c5980fb] Aug 13 07:20:46.019485 kernel: ACPI: Reserving DBG2 table memory at [mem 0x8c598100-0x8c598153] Aug 13 07:20:46.019490 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c598158-0x8c599cbe] Aug 13 07:20:46.019494 kernel: ACPI: Reserving DMAR table memory at [mem 0x8c599cc0-0x8c599d2f] Aug 13 07:20:46.019499 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599d30-0x8c599e73] Aug 13 07:20:46.019504 kernel: ACPI: Reserving TPM2 table memory at [mem 0x8c599e78-0x8c599eab] Aug 13 07:20:46.019510 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599eb0-0x8c59ac3e] Aug 13 07:20:46.019515 kernel: ACPI: Reserving WSMT table memory at [mem 0x8c59ac40-0x8c59ac67] Aug 13 07:20:46.019520 kernel: ACPI: Reserving EINJ table memory at [mem 0x8c59ac68-0x8c59ad97] Aug 13 07:20:46.019525 kernel: ACPI: Reserving ERST table memory at [mem 0x8c59ad98-0x8c59afc7] Aug 13 07:20:46.019530 kernel: ACPI: Reserving BERT table memory at [mem 0x8c59afc8-0x8c59aff7] Aug 13 07:20:46.019534 kernel: ACPI: Reserving HEST table memory at [mem 0x8c59aff8-0x8c59b273] Aug 13 07:20:46.019539 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c59b278-0x8c59b3d9] Aug 13 07:20:46.019544 kernel: No NUMA configuration found Aug 13 07:20:46.019549 kernel: Faking a node at [mem 0x0000000000000000-0x000000086effffff] Aug 13 07:20:46.019554 kernel: NODE_DATA(0) allocated [mem 0x86effa000-0x86effffff] Aug 13 07:20:46.019560 kernel: Zone ranges: Aug 13 07:20:46.019565 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Aug 13 07:20:46.019570 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Aug 13 07:20:46.019575 kernel: Normal [mem 0x0000000100000000-0x000000086effffff] Aug 13 07:20:46.019580 kernel: Movable zone start for each node Aug 13 07:20:46.019585 kernel: Early memory node ranges Aug 13 07:20:46.019590 kernel: node 0: [mem 0x0000000000001000-0x0000000000098fff] Aug 13 07:20:46.019595 kernel: node 0: [mem 0x0000000000100000-0x000000003fffffff] Aug 13 07:20:46.019600 kernel: node 0: [mem 0x0000000040400000-0x00000000819c6fff] Aug 13 07:20:46.019606 kernel: node 0: [mem 0x00000000819c9000-0x000000008afcdfff] Aug 13 07:20:46.019611 kernel: node 0: [mem 0x000000008c0b3000-0x000000008c23bfff] Aug 13 07:20:46.019615 kernel: node 0: [mem 0x000000008eeff000-0x000000008eefffff] Aug 13 07:20:46.019621 kernel: node 0: [mem 0x0000000100000000-0x000000086effffff] Aug 13 07:20:46.019630 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000086effffff] Aug 13 07:20:46.019635 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Aug 13 07:20:46.019640 kernel: On node 0, zone DMA: 103 pages in unavailable ranges Aug 13 07:20:46.019645 kernel: On node 0, zone DMA32: 1024 pages in unavailable ranges Aug 13 07:20:46.019652 kernel: On node 0, zone DMA32: 2 pages in unavailable ranges Aug 13 07:20:46.019657 kernel: On node 0, zone DMA32: 4325 pages in unavailable ranges Aug 13 07:20:46.019662 kernel: On node 0, zone DMA32: 11459 pages in unavailable ranges Aug 13 07:20:46.019667 kernel: On node 0, zone Normal: 4352 pages in unavailable ranges Aug 13 07:20:46.019673 kernel: On node 0, zone Normal: 4096 pages in unavailable ranges Aug 13 07:20:46.019678 kernel: ACPI: PM-Timer IO Port: 0x1808 Aug 13 07:20:46.019683 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Aug 13 07:20:46.019689 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Aug 13 07:20:46.019694 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Aug 13 07:20:46.019700 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Aug 13 07:20:46.019705 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Aug 13 07:20:46.019710 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Aug 13 07:20:46.019716 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Aug 13 07:20:46.019721 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Aug 13 07:20:46.019726 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Aug 13 07:20:46.019731 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Aug 13 07:20:46.019736 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Aug 13 07:20:46.019741 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Aug 13 07:20:46.019748 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Aug 13 07:20:46.019753 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Aug 13 07:20:46.019758 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Aug 13 07:20:46.019763 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Aug 13 07:20:46.019768 kernel: IOAPIC[0]: apic_id 2, version 32, address 0xfec00000, GSI 0-119 Aug 13 07:20:46.019774 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Aug 13 07:20:46.019779 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Aug 13 07:20:46.019784 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Aug 13 07:20:46.019789 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Aug 13 07:20:46.019796 kernel: TSC deadline timer available Aug 13 07:20:46.019801 kernel: smpboot: Allowing 16 CPUs, 0 hotplug CPUs Aug 13 07:20:46.019806 kernel: [mem 0x90000000-0xdfffffff] available for PCI devices Aug 13 07:20:46.019812 kernel: Booting paravirtualized kernel on bare hardware Aug 13 07:20:46.019817 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Aug 13 07:20:46.019823 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Aug 13 07:20:46.019828 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u262144 Aug 13 07:20:46.019833 kernel: pcpu-alloc: s197096 r8192 d32280 u262144 alloc=1*2097152 Aug 13 07:20:46.019838 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Aug 13 07:20:46.019845 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:20:46.019851 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 07:20:46.019856 kernel: random: crng init done Aug 13 07:20:46.019861 kernel: Dentry cache hash table entries: 4194304 (order: 13, 33554432 bytes, linear) Aug 13 07:20:46.019866 kernel: Inode-cache hash table entries: 2097152 (order: 12, 16777216 bytes, linear) Aug 13 07:20:46.019872 kernel: Fallback order for Node 0: 0 Aug 13 07:20:46.019877 kernel: Built 1 zonelists, mobility grouping on. Total pages: 8232416 Aug 13 07:20:46.019882 kernel: Policy zone: Normal Aug 13 07:20:46.019888 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 07:20:46.019894 kernel: software IO TLB: area num 16. Aug 13 07:20:46.019899 kernel: Memory: 32720316K/33452984K available (12288K kernel code, 2295K rwdata, 22748K rodata, 42876K init, 2316K bss, 732408K reserved, 0K cma-reserved) Aug 13 07:20:46.019904 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Aug 13 07:20:46.019910 kernel: ftrace: allocating 37968 entries in 149 pages Aug 13 07:20:46.019915 kernel: ftrace: allocated 149 pages with 4 groups Aug 13 07:20:46.019920 kernel: Dynamic Preempt: voluntary Aug 13 07:20:46.019926 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 07:20:46.019931 kernel: rcu: RCU event tracing is enabled. Aug 13 07:20:46.019938 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Aug 13 07:20:46.019943 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 07:20:46.019948 kernel: Rude variant of Tasks RCU enabled. Aug 13 07:20:46.019954 kernel: Tracing variant of Tasks RCU enabled. Aug 13 07:20:46.019959 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 07:20:46.019964 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Aug 13 07:20:46.019970 kernel: NR_IRQS: 33024, nr_irqs: 2184, preallocated irqs: 16 Aug 13 07:20:46.019975 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 07:20:46.019980 kernel: Console: colour dummy device 80x25 Aug 13 07:20:46.019985 kernel: printk: console [tty0] enabled Aug 13 07:20:46.019991 kernel: printk: console [ttyS1] enabled Aug 13 07:20:46.019997 kernel: ACPI: Core revision 20230628 Aug 13 07:20:46.020002 kernel: hpet: HPET dysfunctional in PC10. Force disabled. Aug 13 07:20:46.020007 kernel: APIC: Switch to symmetric I/O mode setup Aug 13 07:20:46.020012 kernel: DMAR: Host address width 39 Aug 13 07:20:46.020018 kernel: DMAR: DRHD base: 0x000000fed91000 flags: 0x1 Aug 13 07:20:46.020023 kernel: DMAR: dmar0: reg_base_addr fed91000 ver 1:0 cap d2008c40660462 ecap f050da Aug 13 07:20:46.020029 kernel: DMAR: RMRR base: 0x0000008cf19000 end: 0x0000008d162fff Aug 13 07:20:46.020034 kernel: DMAR-IR: IOAPIC id 2 under DRHD base 0xfed91000 IOMMU 0 Aug 13 07:20:46.020040 kernel: DMAR-IR: HPET id 0 under DRHD base 0xfed91000 Aug 13 07:20:46.020045 kernel: DMAR-IR: Queued invalidation will be enabled to support x2apic and Intr-remapping. Aug 13 07:20:46.020050 kernel: DMAR-IR: Enabled IRQ remapping in x2apic mode Aug 13 07:20:46.020056 kernel: x2apic enabled Aug 13 07:20:46.020061 kernel: APIC: Switched APIC routing to: cluster x2apic Aug 13 07:20:46.020068 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x3101f59f5e6, max_idle_ns: 440795259996 ns Aug 13 07:20:46.020074 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 6799.81 BogoMIPS (lpj=3399906) Aug 13 07:20:46.020098 kernel: CPU0: Thermal monitoring enabled (TM1) Aug 13 07:20:46.020103 kernel: process: using mwait in idle threads Aug 13 07:20:46.020110 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Aug 13 07:20:46.020115 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Aug 13 07:20:46.020134 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Aug 13 07:20:46.020139 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Aug 13 07:20:46.020144 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Aug 13 07:20:46.020149 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Aug 13 07:20:46.020155 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Aug 13 07:20:46.020160 kernel: RETBleed: Mitigation: Enhanced IBRS Aug 13 07:20:46.020165 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Aug 13 07:20:46.020170 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Aug 13 07:20:46.020175 kernel: TAA: Mitigation: TSX disabled Aug 13 07:20:46.020181 kernel: MMIO Stale Data: Mitigation: Clear CPU buffers Aug 13 07:20:46.020187 kernel: SRBDS: Mitigation: Microcode Aug 13 07:20:46.020192 kernel: GDS: Mitigation: Microcode Aug 13 07:20:46.020197 kernel: ITS: Mitigation: Aligned branch/return thunks Aug 13 07:20:46.020202 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Aug 13 07:20:46.020207 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Aug 13 07:20:46.020213 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Aug 13 07:20:46.020218 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Aug 13 07:20:46.020223 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Aug 13 07:20:46.020228 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Aug 13 07:20:46.020233 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Aug 13 07:20:46.020240 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Aug 13 07:20:46.020245 kernel: x86/fpu: Enabled xstate features 0x1f, context size is 960 bytes, using 'compacted' format. Aug 13 07:20:46.020250 kernel: Freeing SMP alternatives memory: 32K Aug 13 07:20:46.020255 kernel: pid_max: default: 32768 minimum: 301 Aug 13 07:20:46.020261 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Aug 13 07:20:46.020266 kernel: landlock: Up and running. Aug 13 07:20:46.020271 kernel: SELinux: Initializing. Aug 13 07:20:46.020276 kernel: Mount-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.020281 kernel: Mountpoint-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.020287 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Aug 13 07:20:46.020292 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Aug 13 07:20:46.020298 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Aug 13 07:20:46.020304 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Aug 13 07:20:46.020309 kernel: Performance Events: PEBS fmt3+, Skylake events, 32-deep LBR, full-width counters, Intel PMU driver. Aug 13 07:20:46.020314 kernel: ... version: 4 Aug 13 07:20:46.020320 kernel: ... bit width: 48 Aug 13 07:20:46.020325 kernel: ... generic registers: 4 Aug 13 07:20:46.020330 kernel: ... value mask: 0000ffffffffffff Aug 13 07:20:46.020335 kernel: ... max period: 00007fffffffffff Aug 13 07:20:46.020341 kernel: ... fixed-purpose events: 3 Aug 13 07:20:46.020347 kernel: ... event mask: 000000070000000f Aug 13 07:20:46.020352 kernel: signal: max sigframe size: 2032 Aug 13 07:20:46.020357 kernel: Estimated ratio of average max frequency by base frequency (times 1024): 1445 Aug 13 07:20:46.020363 kernel: rcu: Hierarchical SRCU implementation. Aug 13 07:20:46.020368 kernel: rcu: Max phase no-delay instances is 400. Aug 13 07:20:46.020373 kernel: NMI watchdog: Enabled. Permanently consumes one hw-PMU counter. Aug 13 07:20:46.020379 kernel: smp: Bringing up secondary CPUs ... Aug 13 07:20:46.020384 kernel: smpboot: x86: Booting SMP configuration: Aug 13 07:20:46.020389 kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6 #7 #8 #9 #10 #11 #12 #13 #14 #15 Aug 13 07:20:46.020396 kernel: Transient Scheduler Attacks: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Aug 13 07:20:46.020401 kernel: smp: Brought up 1 node, 16 CPUs Aug 13 07:20:46.020406 kernel: smpboot: Max logical packages: 1 Aug 13 07:20:46.020412 kernel: smpboot: Total of 16 processors activated (108796.99 BogoMIPS) Aug 13 07:20:46.020417 kernel: devtmpfs: initialized Aug 13 07:20:46.020422 kernel: x86/mm: Memory block size: 128MB Aug 13 07:20:46.020427 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x819c7000-0x819c7fff] (4096 bytes) Aug 13 07:20:46.020433 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x8c23c000-0x8c66dfff] (4399104 bytes) Aug 13 07:20:46.020438 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 07:20:46.020444 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Aug 13 07:20:46.020450 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 07:20:46.020455 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 07:20:46.020460 kernel: audit: initializing netlink subsys (disabled) Aug 13 07:20:46.020465 kernel: audit: type=2000 audit(1755069640.039:1): state=initialized audit_enabled=0 res=1 Aug 13 07:20:46.020470 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 07:20:46.020476 kernel: thermal_sys: Registered thermal governor 'user_space' Aug 13 07:20:46.020481 kernel: cpuidle: using governor menu Aug 13 07:20:46.020486 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 07:20:46.020492 kernel: dca service started, version 1.12.1 Aug 13 07:20:46.020498 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Aug 13 07:20:46.020503 kernel: PCI: Using configuration type 1 for base access Aug 13 07:20:46.020508 kernel: ENERGY_PERF_BIAS: Set to 'normal', was 'performance' Aug 13 07:20:46.020513 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Aug 13 07:20:46.020519 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 07:20:46.020524 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 07:20:46.020529 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 07:20:46.020535 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 07:20:46.020540 kernel: ACPI: Added _OSI(Module Device) Aug 13 07:20:46.020545 kernel: ACPI: Added _OSI(Processor Device) Aug 13 07:20:46.020551 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 07:20:46.020556 kernel: ACPI: 12 ACPI AML tables successfully acquired and loaded Aug 13 07:20:46.020561 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020566 kernel: ACPI: SSDT 0xFFFF9BA581AF5400 000400 (v02 PmRef Cpu0Cst 00003001 INTL 20160527) Aug 13 07:20:46.020572 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020577 kernel: ACPI: SSDT 0xFFFF9BA581AEF800 000683 (v02 PmRef Cpu0Ist 00003000 INTL 20160527) Aug 13 07:20:46.020582 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020588 kernel: ACPI: SSDT 0xFFFF9BA580247E00 0000F4 (v02 PmRef Cpu0Psd 00003000 INTL 20160527) Aug 13 07:20:46.020594 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020599 kernel: ACPI: SSDT 0xFFFF9BA581AEB800 0005FC (v02 PmRef ApIst 00003000 INTL 20160527) Aug 13 07:20:46.020604 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020609 kernel: ACPI: SSDT 0xFFFF9BA58012D000 000AB0 (v02 PmRef ApPsd 00003000 INTL 20160527) Aug 13 07:20:46.020614 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020620 kernel: ACPI: SSDT 0xFFFF9BA581AF0800 00030A (v02 PmRef ApCst 00003000 INTL 20160527) Aug 13 07:20:46.020625 kernel: ACPI: _OSC evaluated successfully for all CPUs Aug 13 07:20:46.020630 kernel: ACPI: Interpreter enabled Aug 13 07:20:46.020636 kernel: ACPI: PM: (supports S0 S5) Aug 13 07:20:46.020642 kernel: ACPI: Using IOAPIC for interrupt routing Aug 13 07:20:46.020647 kernel: HEST: Enabling Firmware First mode for corrected errors. Aug 13 07:20:46.020652 kernel: mce: [Firmware Bug]: Ignoring request to disable invalid MCA bank 14. Aug 13 07:20:46.020657 kernel: HEST: Table parsing has been initialized. Aug 13 07:20:46.020663 kernel: GHES: APEI firmware first mode is enabled by APEI bit and WHEA _OSC. Aug 13 07:20:46.020668 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Aug 13 07:20:46.020673 kernel: PCI: Ignoring E820 reservations for host bridge windows Aug 13 07:20:46.020678 kernel: ACPI: Enabled 9 GPEs in block 00 to 7F Aug 13 07:20:46.020685 kernel: ACPI: \_SB_.PCI0.XDCI.USBC: New power resource Aug 13 07:20:46.020690 kernel: ACPI: \_SB_.PCI0.SAT0.VOL0.V0PR: New power resource Aug 13 07:20:46.020695 kernel: ACPI: \_SB_.PCI0.SAT0.VOL1.V1PR: New power resource Aug 13 07:20:46.020701 kernel: ACPI: \_SB_.PCI0.SAT0.VOL2.V2PR: New power resource Aug 13 07:20:46.020706 kernel: ACPI: \_SB_.PCI0.CNVW.WRST: New power resource Aug 13 07:20:46.020711 kernel: ACPI: \_TZ_.FN00: New power resource Aug 13 07:20:46.020717 kernel: ACPI: \_TZ_.FN01: New power resource Aug 13 07:20:46.020722 kernel: ACPI: \_TZ_.FN02: New power resource Aug 13 07:20:46.020727 kernel: ACPI: \_TZ_.FN03: New power resource Aug 13 07:20:46.020733 kernel: ACPI: \_TZ_.FN04: New power resource Aug 13 07:20:46.020739 kernel: ACPI: \PIN_: New power resource Aug 13 07:20:46.020744 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-fe]) Aug 13 07:20:46.020818 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 07:20:46.020873 kernel: acpi PNP0A08:00: _OSC: platform does not support [AER] Aug 13 07:20:46.020923 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability LTR] Aug 13 07:20:46.020931 kernel: PCI host bridge to bus 0000:00 Aug 13 07:20:46.020983 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Aug 13 07:20:46.021031 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Aug 13 07:20:46.021077 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Aug 13 07:20:46.021158 kernel: pci_bus 0000:00: root bus resource [mem 0x90000000-0xdfffffff window] Aug 13 07:20:46.021200 kernel: pci_bus 0000:00: root bus resource [mem 0xfc800000-0xfe7fffff window] Aug 13 07:20:46.021244 kernel: pci_bus 0000:00: root bus resource [bus 00-fe] Aug 13 07:20:46.021302 kernel: pci 0000:00:00.0: [8086:3e31] type 00 class 0x060000 Aug 13 07:20:46.021361 kernel: pci 0000:00:01.0: [8086:1901] type 01 class 0x060400 Aug 13 07:20:46.021413 kernel: pci 0000:00:01.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.021468 kernel: pci 0000:00:08.0: [8086:1911] type 00 class 0x088000 Aug 13 07:20:46.021518 kernel: pci 0000:00:08.0: reg 0x10: [mem 0x9551f000-0x9551ffff 64bit] Aug 13 07:20:46.021571 kernel: pci 0000:00:12.0: [8086:a379] type 00 class 0x118000 Aug 13 07:20:46.021621 kernel: pci 0000:00:12.0: reg 0x10: [mem 0x9551e000-0x9551efff 64bit] Aug 13 07:20:46.021679 kernel: pci 0000:00:14.0: [8086:a36d] type 00 class 0x0c0330 Aug 13 07:20:46.021729 kernel: pci 0000:00:14.0: reg 0x10: [mem 0x95500000-0x9550ffff 64bit] Aug 13 07:20:46.021778 kernel: pci 0000:00:14.0: PME# supported from D3hot D3cold Aug 13 07:20:46.021832 kernel: pci 0000:00:14.2: [8086:a36f] type 00 class 0x050000 Aug 13 07:20:46.021881 kernel: pci 0000:00:14.2: reg 0x10: [mem 0x95512000-0x95513fff 64bit] Aug 13 07:20:46.021929 kernel: pci 0000:00:14.2: reg 0x18: [mem 0x9551d000-0x9551dfff 64bit] Aug 13 07:20:46.021982 kernel: pci 0000:00:15.0: [8086:a368] type 00 class 0x0c8000 Aug 13 07:20:46.022035 kernel: pci 0000:00:15.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Aug 13 07:20:46.022120 kernel: pci 0000:00:15.1: [8086:a369] type 00 class 0x0c8000 Aug 13 07:20:46.022170 kernel: pci 0000:00:15.1: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Aug 13 07:20:46.022223 kernel: pci 0000:00:16.0: [8086:a360] type 00 class 0x078000 Aug 13 07:20:46.022272 kernel: pci 0000:00:16.0: reg 0x10: [mem 0x9551a000-0x9551afff 64bit] Aug 13 07:20:46.022325 kernel: pci 0000:00:16.0: PME# supported from D3hot Aug 13 07:20:46.022377 kernel: pci 0000:00:16.1: [8086:a361] type 00 class 0x078000 Aug 13 07:20:46.022426 kernel: pci 0000:00:16.1: reg 0x10: [mem 0x95519000-0x95519fff 64bit] Aug 13 07:20:46.022484 kernel: pci 0000:00:16.1: PME# supported from D3hot Aug 13 07:20:46.022538 kernel: pci 0000:00:16.4: [8086:a364] type 00 class 0x078000 Aug 13 07:20:46.022587 kernel: pci 0000:00:16.4: reg 0x10: [mem 0x95518000-0x95518fff 64bit] Aug 13 07:20:46.022636 kernel: pci 0000:00:16.4: PME# supported from D3hot Aug 13 07:20:46.022693 kernel: pci 0000:00:17.0: [8086:a352] type 00 class 0x010601 Aug 13 07:20:46.022743 kernel: pci 0000:00:17.0: reg 0x10: [mem 0x95510000-0x95511fff] Aug 13 07:20:46.022791 kernel: pci 0000:00:17.0: reg 0x14: [mem 0x95517000-0x955170ff] Aug 13 07:20:46.022840 kernel: pci 0000:00:17.0: reg 0x18: [io 0x6050-0x6057] Aug 13 07:20:46.022888 kernel: pci 0000:00:17.0: reg 0x1c: [io 0x6040-0x6043] Aug 13 07:20:46.022938 kernel: pci 0000:00:17.0: reg 0x20: [io 0x6020-0x603f] Aug 13 07:20:46.022986 kernel: pci 0000:00:17.0: reg 0x24: [mem 0x95516000-0x955167ff] Aug 13 07:20:46.023038 kernel: pci 0000:00:17.0: PME# supported from D3hot Aug 13 07:20:46.023134 kernel: pci 0000:00:1b.0: [8086:a340] type 01 class 0x060400 Aug 13 07:20:46.023187 kernel: pci 0000:00:1b.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023242 kernel: pci 0000:00:1b.4: [8086:a32c] type 01 class 0x060400 Aug 13 07:20:46.023291 kernel: pci 0000:00:1b.4: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023347 kernel: pci 0000:00:1b.5: [8086:a32d] type 01 class 0x060400 Aug 13 07:20:46.023397 kernel: pci 0000:00:1b.5: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023450 kernel: pci 0000:00:1c.0: [8086:a338] type 01 class 0x060400 Aug 13 07:20:46.023500 kernel: pci 0000:00:1c.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023555 kernel: pci 0000:00:1c.3: [8086:a33b] type 01 class 0x060400 Aug 13 07:20:46.023607 kernel: pci 0000:00:1c.3: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023663 kernel: pci 0000:00:1e.0: [8086:a328] type 00 class 0x078000 Aug 13 07:20:46.023712 kernel: pci 0000:00:1e.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Aug 13 07:20:46.023766 kernel: pci 0000:00:1f.0: [8086:a309] type 00 class 0x060100 Aug 13 07:20:46.023818 kernel: pci 0000:00:1f.4: [8086:a323] type 00 class 0x0c0500 Aug 13 07:20:46.023869 kernel: pci 0000:00:1f.4: reg 0x10: [mem 0x95514000-0x955140ff 64bit] Aug 13 07:20:46.023920 kernel: pci 0000:00:1f.4: reg 0x20: [io 0xefa0-0xefbf] Aug 13 07:20:46.023974 kernel: pci 0000:00:1f.5: [8086:a324] type 00 class 0x0c8000 Aug 13 07:20:46.024023 kernel: pci 0000:00:1f.5: reg 0x10: [mem 0xfe010000-0xfe010fff] Aug 13 07:20:46.024105 kernel: pci 0000:01:00.0: [15b3:1015] type 00 class 0x020000 Aug 13 07:20:46.024174 kernel: pci 0000:01:00.0: reg 0x10: [mem 0x92000000-0x93ffffff 64bit pref] Aug 13 07:20:46.024225 kernel: pci 0000:01:00.0: reg 0x30: [mem 0x95200000-0x952fffff pref] Aug 13 07:20:46.024275 kernel: pci 0000:01:00.0: PME# supported from D3cold Aug 13 07:20:46.024329 kernel: pci 0000:01:00.0: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Aug 13 07:20:46.024380 kernel: pci 0000:01:00.0: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 07:20:46.024437 kernel: pci 0000:01:00.1: [15b3:1015] type 00 class 0x020000 Aug 13 07:20:46.024488 kernel: pci 0000:01:00.1: reg 0x10: [mem 0x90000000-0x91ffffff 64bit pref] Aug 13 07:20:46.024540 kernel: pci 0000:01:00.1: reg 0x30: [mem 0x95100000-0x951fffff pref] Aug 13 07:20:46.024590 kernel: pci 0000:01:00.1: PME# supported from D3cold Aug 13 07:20:46.024641 kernel: pci 0000:01:00.1: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Aug 13 07:20:46.024694 kernel: pci 0000:01:00.1: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 07:20:46.024745 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Aug 13 07:20:46.024794 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Aug 13 07:20:46.024844 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Aug 13 07:20:46.024892 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Aug 13 07:20:46.024950 kernel: pci 0000:03:00.0: working around ROM BAR overlap defect Aug 13 07:20:46.025002 kernel: pci 0000:03:00.0: [8086:1533] type 00 class 0x020000 Aug 13 07:20:46.025055 kernel: pci 0000:03:00.0: reg 0x10: [mem 0x95400000-0x9547ffff] Aug 13 07:20:46.025141 kernel: pci 0000:03:00.0: reg 0x18: [io 0x5000-0x501f] Aug 13 07:20:46.025191 kernel: pci 0000:03:00.0: reg 0x1c: [mem 0x95480000-0x95483fff] Aug 13 07:20:46.025241 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.025291 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Aug 13 07:20:46.025341 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Aug 13 07:20:46.025389 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Aug 13 07:20:46.025445 kernel: pci 0000:04:00.0: working around ROM BAR overlap defect Aug 13 07:20:46.025500 kernel: pci 0000:04:00.0: [8086:1533] type 00 class 0x020000 Aug 13 07:20:46.025550 kernel: pci 0000:04:00.0: reg 0x10: [mem 0x95300000-0x9537ffff] Aug 13 07:20:46.025600 kernel: pci 0000:04:00.0: reg 0x18: [io 0x4000-0x401f] Aug 13 07:20:46.025649 kernel: pci 0000:04:00.0: reg 0x1c: [mem 0x95380000-0x95383fff] Aug 13 07:20:46.025700 kernel: pci 0000:04:00.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.025750 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Aug 13 07:20:46.025800 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Aug 13 07:20:46.025851 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Aug 13 07:20:46.025901 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Aug 13 07:20:46.025957 kernel: pci 0000:06:00.0: [1a03:1150] type 01 class 0x060400 Aug 13 07:20:46.026008 kernel: pci 0000:06:00.0: enabling Extended Tags Aug 13 07:20:46.026058 kernel: pci 0000:06:00.0: supports D1 D2 Aug 13 07:20:46.026151 kernel: pci 0000:06:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 07:20:46.026202 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Aug 13 07:20:46.026253 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.026302 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.026354 kernel: pci_bus 0000:07: extended config space not accessible Aug 13 07:20:46.026412 kernel: pci 0000:07:00.0: [1a03:2000] type 00 class 0x030000 Aug 13 07:20:46.026464 kernel: pci 0000:07:00.0: reg 0x10: [mem 0x94000000-0x94ffffff] Aug 13 07:20:46.026517 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x95000000-0x9501ffff] Aug 13 07:20:46.026571 kernel: pci 0000:07:00.0: reg 0x18: [io 0x3000-0x307f] Aug 13 07:20:46.026622 kernel: pci 0000:07:00.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Aug 13 07:20:46.026677 kernel: pci 0000:07:00.0: supports D1 D2 Aug 13 07:20:46.026729 kernel: pci 0000:07:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 07:20:46.026781 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Aug 13 07:20:46.026832 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.026882 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.026891 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 0 Aug 13 07:20:46.026897 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 1 Aug 13 07:20:46.026904 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 0 Aug 13 07:20:46.026910 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 0 Aug 13 07:20:46.026916 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 0 Aug 13 07:20:46.026921 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 0 Aug 13 07:20:46.026927 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 0 Aug 13 07:20:46.026932 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 0 Aug 13 07:20:46.026938 kernel: iommu: Default domain type: Translated Aug 13 07:20:46.026944 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Aug 13 07:20:46.026949 kernel: PCI: Using ACPI for IRQ routing Aug 13 07:20:46.026956 kernel: PCI: pci_cache_line_size set to 64 bytes Aug 13 07:20:46.026961 kernel: e820: reserve RAM buffer [mem 0x00099800-0x0009ffff] Aug 13 07:20:46.026967 kernel: e820: reserve RAM buffer [mem 0x819c7000-0x83ffffff] Aug 13 07:20:46.026973 kernel: e820: reserve RAM buffer [mem 0x8afce000-0x8bffffff] Aug 13 07:20:46.026979 kernel: e820: reserve RAM buffer [mem 0x8c23c000-0x8fffffff] Aug 13 07:20:46.026984 kernel: e820: reserve RAM buffer [mem 0x8ef00000-0x8fffffff] Aug 13 07:20:46.026990 kernel: e820: reserve RAM buffer [mem 0x86f000000-0x86fffffff] Aug 13 07:20:46.027042 kernel: pci 0000:07:00.0: vgaarb: setting as boot VGA device Aug 13 07:20:46.027122 kernel: pci 0000:07:00.0: vgaarb: bridge control possible Aug 13 07:20:46.027197 kernel: pci 0000:07:00.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Aug 13 07:20:46.027206 kernel: vgaarb: loaded Aug 13 07:20:46.027212 kernel: clocksource: Switched to clocksource tsc-early Aug 13 07:20:46.027217 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 07:20:46.027223 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 07:20:46.027229 kernel: pnp: PnP ACPI init Aug 13 07:20:46.027281 kernel: system 00:00: [mem 0x40000000-0x403fffff] has been reserved Aug 13 07:20:46.027332 kernel: pnp 00:02: [dma 0 disabled] Aug 13 07:20:46.027383 kernel: pnp 00:03: [dma 0 disabled] Aug 13 07:20:46.027432 kernel: system 00:04: [io 0x0680-0x069f] has been reserved Aug 13 07:20:46.027477 kernel: system 00:04: [io 0x164e-0x164f] has been reserved Aug 13 07:20:46.027526 kernel: system 00:05: [mem 0xfed10000-0xfed17fff] has been reserved Aug 13 07:20:46.027571 kernel: system 00:05: [mem 0xfed18000-0xfed18fff] has been reserved Aug 13 07:20:46.027615 kernel: system 00:05: [mem 0xfed19000-0xfed19fff] has been reserved Aug 13 07:20:46.027663 kernel: system 00:05: [mem 0xe0000000-0xefffffff] has been reserved Aug 13 07:20:46.027707 kernel: system 00:05: [mem 0xfed20000-0xfed3ffff] has been reserved Aug 13 07:20:46.027752 kernel: system 00:05: [mem 0xfed90000-0xfed93fff] could not be reserved Aug 13 07:20:46.027796 kernel: system 00:05: [mem 0xfed45000-0xfed8ffff] has been reserved Aug 13 07:20:46.027842 kernel: system 00:05: [mem 0xfee00000-0xfeefffff] could not be reserved Aug 13 07:20:46.027891 kernel: system 00:06: [io 0x1800-0x18fe] could not be reserved Aug 13 07:20:46.027940 kernel: system 00:06: [mem 0xfd000000-0xfd69ffff] has been reserved Aug 13 07:20:46.027987 kernel: system 00:06: [mem 0xfd6c0000-0xfd6cffff] has been reserved Aug 13 07:20:46.028032 kernel: system 00:06: [mem 0xfd6f0000-0xfdffffff] has been reserved Aug 13 07:20:46.028099 kernel: system 00:06: [mem 0xfe000000-0xfe01ffff] could not be reserved Aug 13 07:20:46.028163 kernel: system 00:06: [mem 0xfe200000-0xfe7fffff] has been reserved Aug 13 07:20:46.028208 kernel: system 00:06: [mem 0xff000000-0xffffffff] has been reserved Aug 13 07:20:46.028256 kernel: system 00:07: [io 0x2000-0x20fe] has been reserved Aug 13 07:20:46.028265 kernel: pnp: PnP ACPI: found 9 devices Aug 13 07:20:46.028273 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Aug 13 07:20:46.028279 kernel: NET: Registered PF_INET protocol family Aug 13 07:20:46.028284 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 07:20:46.028290 kernel: tcp_listen_portaddr_hash hash table entries: 16384 (order: 6, 262144 bytes, linear) Aug 13 07:20:46.028296 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 07:20:46.028302 kernel: TCP established hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 07:20:46.028307 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 07:20:46.028313 kernel: TCP: Hash tables configured (established 262144 bind 65536) Aug 13 07:20:46.028318 kernel: UDP hash table entries: 16384 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.028325 kernel: UDP-Lite hash table entries: 16384 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.028331 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 07:20:46.028337 kernel: NET: Registered PF_XDP protocol family Aug 13 07:20:46.028387 kernel: pci 0000:00:15.0: BAR 0: assigned [mem 0x95515000-0x95515fff 64bit] Aug 13 07:20:46.028437 kernel: pci 0000:00:15.1: BAR 0: assigned [mem 0x9551b000-0x9551bfff 64bit] Aug 13 07:20:46.028487 kernel: pci 0000:00:1e.0: BAR 0: assigned [mem 0x9551c000-0x9551cfff 64bit] Aug 13 07:20:46.028539 kernel: pci 0000:01:00.0: BAR 7: no space for [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028590 kernel: pci 0000:01:00.0: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028643 kernel: pci 0000:01:00.1: BAR 7: no space for [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028695 kernel: pci 0000:01:00.1: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028744 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Aug 13 07:20:46.028794 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Aug 13 07:20:46.028843 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Aug 13 07:20:46.028893 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Aug 13 07:20:46.028946 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Aug 13 07:20:46.028995 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Aug 13 07:20:46.029047 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Aug 13 07:20:46.029134 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Aug 13 07:20:46.029185 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Aug 13 07:20:46.029233 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Aug 13 07:20:46.029283 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Aug 13 07:20:46.029336 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Aug 13 07:20:46.029387 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.029437 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.029487 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Aug 13 07:20:46.029537 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.029586 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.029632 kernel: pci_bus 0000:00: Some PCI device resources are unassigned, try booting with pci=realloc Aug 13 07:20:46.029675 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Aug 13 07:20:46.029719 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Aug 13 07:20:46.029764 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Aug 13 07:20:46.029807 kernel: pci_bus 0000:00: resource 7 [mem 0x90000000-0xdfffffff window] Aug 13 07:20:46.029849 kernel: pci_bus 0000:00: resource 8 [mem 0xfc800000-0xfe7fffff window] Aug 13 07:20:46.029902 kernel: pci_bus 0000:01: resource 1 [mem 0x95100000-0x952fffff] Aug 13 07:20:46.029947 kernel: pci_bus 0000:01: resource 2 [mem 0x90000000-0x93ffffff 64bit pref] Aug 13 07:20:46.029999 kernel: pci_bus 0000:03: resource 0 [io 0x5000-0x5fff] Aug 13 07:20:46.030047 kernel: pci_bus 0000:03: resource 1 [mem 0x95400000-0x954fffff] Aug 13 07:20:46.030143 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Aug 13 07:20:46.030188 kernel: pci_bus 0000:04: resource 1 [mem 0x95300000-0x953fffff] Aug 13 07:20:46.030236 kernel: pci_bus 0000:06: resource 0 [io 0x3000-0x3fff] Aug 13 07:20:46.030281 kernel: pci_bus 0000:06: resource 1 [mem 0x94000000-0x950fffff] Aug 13 07:20:46.030329 kernel: pci_bus 0000:07: resource 0 [io 0x3000-0x3fff] Aug 13 07:20:46.030375 kernel: pci_bus 0000:07: resource 1 [mem 0x94000000-0x950fffff] Aug 13 07:20:46.030385 kernel: PCI: CLS 64 bytes, default 64 Aug 13 07:20:46.030391 kernel: DMAR: No ATSR found Aug 13 07:20:46.030396 kernel: DMAR: No SATC found Aug 13 07:20:46.030402 kernel: DMAR: dmar0: Using Queued invalidation Aug 13 07:20:46.030452 kernel: pci 0000:00:00.0: Adding to iommu group 0 Aug 13 07:20:46.030501 kernel: pci 0000:00:01.0: Adding to iommu group 1 Aug 13 07:20:46.030552 kernel: pci 0000:00:08.0: Adding to iommu group 2 Aug 13 07:20:46.030602 kernel: pci 0000:00:12.0: Adding to iommu group 3 Aug 13 07:20:46.030651 kernel: pci 0000:00:14.0: Adding to iommu group 4 Aug 13 07:20:46.030704 kernel: pci 0000:00:14.2: Adding to iommu group 4 Aug 13 07:20:46.030752 kernel: pci 0000:00:15.0: Adding to iommu group 5 Aug 13 07:20:46.030801 kernel: pci 0000:00:15.1: Adding to iommu group 5 Aug 13 07:20:46.030849 kernel: pci 0000:00:16.0: Adding to iommu group 6 Aug 13 07:20:46.030899 kernel: pci 0000:00:16.1: Adding to iommu group 6 Aug 13 07:20:46.030947 kernel: pci 0000:00:16.4: Adding to iommu group 6 Aug 13 07:20:46.030997 kernel: pci 0000:00:17.0: Adding to iommu group 7 Aug 13 07:20:46.031045 kernel: pci 0000:00:1b.0: Adding to iommu group 8 Aug 13 07:20:46.031144 kernel: pci 0000:00:1b.4: Adding to iommu group 9 Aug 13 07:20:46.031193 kernel: pci 0000:00:1b.5: Adding to iommu group 10 Aug 13 07:20:46.031243 kernel: pci 0000:00:1c.0: Adding to iommu group 11 Aug 13 07:20:46.031291 kernel: pci 0000:00:1c.3: Adding to iommu group 12 Aug 13 07:20:46.031340 kernel: pci 0000:00:1e.0: Adding to iommu group 13 Aug 13 07:20:46.031390 kernel: pci 0000:00:1f.0: Adding to iommu group 14 Aug 13 07:20:46.031439 kernel: pci 0000:00:1f.4: Adding to iommu group 14 Aug 13 07:20:46.031488 kernel: pci 0000:00:1f.5: Adding to iommu group 14 Aug 13 07:20:46.031542 kernel: pci 0000:01:00.0: Adding to iommu group 1 Aug 13 07:20:46.031593 kernel: pci 0000:01:00.1: Adding to iommu group 1 Aug 13 07:20:46.031644 kernel: pci 0000:03:00.0: Adding to iommu group 15 Aug 13 07:20:46.031696 kernel: pci 0000:04:00.0: Adding to iommu group 16 Aug 13 07:20:46.031746 kernel: pci 0000:06:00.0: Adding to iommu group 17 Aug 13 07:20:46.031800 kernel: pci 0000:07:00.0: Adding to iommu group 17 Aug 13 07:20:46.031808 kernel: DMAR: Intel(R) Virtualization Technology for Directed I/O Aug 13 07:20:46.031814 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Aug 13 07:20:46.031822 kernel: software IO TLB: mapped [mem 0x0000000086fce000-0x000000008afce000] (64MB) Aug 13 07:20:46.031828 kernel: RAPL PMU: API unit is 2^-32 Joules, 3 fixed counters, 655360 ms ovfl timer Aug 13 07:20:46.031833 kernel: RAPL PMU: hw unit of domain pp0-core 2^-14 Joules Aug 13 07:20:46.031839 kernel: RAPL PMU: hw unit of domain package 2^-14 Joules Aug 13 07:20:46.031845 kernel: RAPL PMU: hw unit of domain dram 2^-14 Joules Aug 13 07:20:46.031899 kernel: platform rtc_cmos: registered platform RTC device (no PNP device found) Aug 13 07:20:46.031908 kernel: Initialise system trusted keyrings Aug 13 07:20:46.031914 kernel: workingset: timestamp_bits=39 max_order=23 bucket_order=0 Aug 13 07:20:46.031921 kernel: Key type asymmetric registered Aug 13 07:20:46.031927 kernel: Asymmetric key parser 'x509' registered Aug 13 07:20:46.031932 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Aug 13 07:20:46.031938 kernel: io scheduler mq-deadline registered Aug 13 07:20:46.031944 kernel: io scheduler kyber registered Aug 13 07:20:46.031949 kernel: io scheduler bfq registered Aug 13 07:20:46.031998 kernel: pcieport 0000:00:01.0: PME: Signaling with IRQ 121 Aug 13 07:20:46.032048 kernel: pcieport 0000:00:1b.0: PME: Signaling with IRQ 122 Aug 13 07:20:46.032146 kernel: pcieport 0000:00:1b.4: PME: Signaling with IRQ 123 Aug 13 07:20:46.032199 kernel: pcieport 0000:00:1b.5: PME: Signaling with IRQ 124 Aug 13 07:20:46.032248 kernel: pcieport 0000:00:1c.0: PME: Signaling with IRQ 125 Aug 13 07:20:46.032298 kernel: pcieport 0000:00:1c.3: PME: Signaling with IRQ 126 Aug 13 07:20:46.032351 kernel: thermal LNXTHERM:00: registered as thermal_zone0 Aug 13 07:20:46.032359 kernel: ACPI: thermal: Thermal Zone [TZ00] (28 C) Aug 13 07:20:46.032365 kernel: ERST: Error Record Serialization Table (ERST) support is initialized. Aug 13 07:20:46.032371 kernel: pstore: Using crash dump compression: deflate Aug 13 07:20:46.032377 kernel: pstore: Registered erst as persistent store backend Aug 13 07:20:46.032384 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Aug 13 07:20:46.032390 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 07:20:46.032395 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Aug 13 07:20:46.032401 kernel: 00:03: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Aug 13 07:20:46.032407 kernel: hpet_acpi_add: no address or irqs in _CRS Aug 13 07:20:46.032456 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x1B, rev-id 16) Aug 13 07:20:46.032465 kernel: i8042: PNP: No PS/2 controller found. Aug 13 07:20:46.032509 kernel: rtc_cmos rtc_cmos: RTC can wake from S4 Aug 13 07:20:46.032557 kernel: rtc_cmos rtc_cmos: registered as rtc0 Aug 13 07:20:46.032601 kernel: rtc_cmos rtc_cmos: setting system clock to 2025-08-13T07:20:44 UTC (1755069644) Aug 13 07:20:46.032647 kernel: rtc_cmos rtc_cmos: alarms up to one month, y3k, 114 bytes nvram Aug 13 07:20:46.032655 kernel: intel_pstate: Intel P-state driver initializing Aug 13 07:20:46.032661 kernel: intel_pstate: Disabling energy efficiency optimization Aug 13 07:20:46.032666 kernel: intel_pstate: HWP enabled Aug 13 07:20:46.032672 kernel: vesafb: mode is 1024x768x8, linelength=1024, pages=0 Aug 13 07:20:46.032678 kernel: vesafb: scrolling: redraw Aug 13 07:20:46.032685 kernel: vesafb: Pseudocolor: size=0:8:8:8, shift=0:0:0:0 Aug 13 07:20:46.032690 kernel: vesafb: framebuffer at 0x94000000, mapped to 0x000000005212a1e2, using 768k, total 768k Aug 13 07:20:46.032696 kernel: Console: switching to colour frame buffer device 128x48 Aug 13 07:20:46.032702 kernel: fb0: VESA VGA frame buffer device Aug 13 07:20:46.032707 kernel: NET: Registered PF_INET6 protocol family Aug 13 07:20:46.032713 kernel: Segment Routing with IPv6 Aug 13 07:20:46.032719 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 07:20:46.032724 kernel: NET: Registered PF_PACKET protocol family Aug 13 07:20:46.032730 kernel: Key type dns_resolver registered Aug 13 07:20:46.032736 kernel: microcode: Current revision: 0x00000102 Aug 13 07:20:46.032742 kernel: microcode: Microcode Update Driver: v2.2. Aug 13 07:20:46.032747 kernel: IPI shorthand broadcast: enabled Aug 13 07:20:46.032753 kernel: sched_clock: Marking stable (1561000725, 1379333307)->(4403206493, -1462872461) Aug 13 07:20:46.032759 kernel: registered taskstats version 1 Aug 13 07:20:46.032764 kernel: Loading compiled-in X.509 certificates Aug 13 07:20:46.032770 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.100-flatcar: 264e720147fa8df9744bb9dc1c08171c0cb20041' Aug 13 07:20:46.032775 kernel: Key type .fscrypt registered Aug 13 07:20:46.032781 kernel: Key type fscrypt-provisioning registered Aug 13 07:20:46.032788 kernel: ima: Allocated hash algorithm: sha1 Aug 13 07:20:46.032793 kernel: ima: No architecture policies found Aug 13 07:20:46.032799 kernel: clk: Disabling unused clocks Aug 13 07:20:46.032805 kernel: Freeing unused kernel image (initmem) memory: 42876K Aug 13 07:20:46.032810 kernel: Write protecting the kernel read-only data: 36864k Aug 13 07:20:46.032816 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Aug 13 07:20:46.032822 kernel: Run /init as init process Aug 13 07:20:46.032827 kernel: with arguments: Aug 13 07:20:46.032833 kernel: /init Aug 13 07:20:46.032840 kernel: with environment: Aug 13 07:20:46.032845 kernel: HOME=/ Aug 13 07:20:46.032851 kernel: TERM=linux Aug 13 07:20:46.032856 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 07:20:46.032863 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 07:20:46.032870 systemd[1]: Detected architecture x86-64. Aug 13 07:20:46.032876 systemd[1]: Running in initrd. Aug 13 07:20:46.032883 systemd[1]: No hostname configured, using default hostname. Aug 13 07:20:46.032889 systemd[1]: Hostname set to . Aug 13 07:20:46.032895 systemd[1]: Initializing machine ID from random generator. Aug 13 07:20:46.032901 systemd[1]: Queued start job for default target initrd.target. Aug 13 07:20:46.032907 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:20:46.032913 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:20:46.032919 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 07:20:46.032925 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:20:46.032932 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 07:20:46.032938 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 07:20:46.032944 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 07:20:46.032951 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 07:20:46.032957 kernel: tsc: Refined TSC clocksource calibration: 3407.999 MHz Aug 13 07:20:46.032963 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd336761, max_idle_ns: 440795243819 ns Aug 13 07:20:46.032968 kernel: clocksource: Switched to clocksource tsc Aug 13 07:20:46.032975 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:20:46.032981 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:20:46.032987 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:20:46.032993 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:20:46.032999 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:20:46.033005 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:20:46.033011 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:20:46.033016 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:20:46.033022 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 07:20:46.033029 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Aug 13 07:20:46.033035 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:20:46.033041 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:20:46.033047 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:20:46.033053 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:20:46.033059 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 07:20:46.033068 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:20:46.033074 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 07:20:46.033101 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 07:20:46.033107 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:20:46.033137 systemd-journald[267]: Collecting audit messages is disabled. Aug 13 07:20:46.033151 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:20:46.033158 systemd-journald[267]: Journal started Aug 13 07:20:46.033172 systemd-journald[267]: Runtime Journal (/run/log/journal/59478b815246437cb8b89f1def936f76) is 8.0M, max 639.9M, 631.9M free. Aug 13 07:20:46.047556 systemd-modules-load[268]: Inserted module 'overlay' Aug 13 07:20:46.068067 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:20:46.088082 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:20:46.100381 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 07:20:46.100550 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:20:46.100690 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 07:20:46.101655 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 07:20:46.101989 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:20:46.143099 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 07:20:46.161650 systemd-modules-load[268]: Inserted module 'br_netfilter' Aug 13 07:20:46.242422 kernel: Bridge firewalling registered Aug 13 07:20:46.161992 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:20:46.259488 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:20:46.280369 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 07:20:46.301503 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:20:46.347291 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:20:46.347733 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:20:46.379348 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:20:46.385187 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:20:46.386445 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:20:46.408046 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:20:46.436369 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 07:20:46.457181 dracut-cmdline[304]: dracut-dracut-053 Aug 13 07:20:46.457181 dracut-cmdline[304]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:20:46.522127 kernel: SCSI subsystem initialized Aug 13 07:20:46.447866 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:20:46.552437 kernel: Loading iSCSI transport class v2.0-870. Aug 13 07:20:46.552450 kernel: iscsi: registered transport (tcp) Aug 13 07:20:46.471364 systemd-resolved[323]: Positive Trust Anchors: Aug 13 07:20:46.582935 kernel: iscsi: registered transport (qla4xxx) Aug 13 07:20:46.471372 systemd-resolved[323]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:20:46.621326 kernel: QLogic iSCSI HBA Driver Aug 13 07:20:46.471403 systemd-resolved[323]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:20:46.473504 systemd-resolved[323]: Defaulting to hostname 'linux'. Aug 13 07:20:46.474205 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:20:46.540172 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:20:46.618099 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 07:20:46.642387 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 07:20:46.821230 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 07:20:46.821255 kernel: device-mapper: uevent: version 1.0.3 Aug 13 07:20:46.840981 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Aug 13 07:20:46.900139 kernel: raid6: avx2x4 gen() 53374 MB/s Aug 13 07:20:46.932141 kernel: raid6: avx2x2 gen() 53939 MB/s Aug 13 07:20:46.968540 kernel: raid6: avx2x1 gen() 45218 MB/s Aug 13 07:20:46.968559 kernel: raid6: using algorithm avx2x2 gen() 53939 MB/s Aug 13 07:20:47.015599 kernel: raid6: .... xor() 31363 MB/s, rmw enabled Aug 13 07:20:47.015617 kernel: raid6: using avx2x2 recovery algorithm Aug 13 07:20:47.057127 kernel: xor: automatically using best checksumming function avx Aug 13 07:20:47.171076 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 07:20:47.176428 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:20:47.206405 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:20:47.213287 systemd-udevd[494]: Using default interface naming scheme 'v255'. Aug 13 07:20:47.217195 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:20:47.258461 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 07:20:47.320245 dracut-pre-trigger[507]: rd.md=0: removing MD RAID activation Aug 13 07:20:47.340880 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:20:47.356418 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:20:47.421579 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:20:47.454334 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 07:20:47.454416 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 07:20:47.480072 kernel: cryptd: max_cpu_qlen set to 1000 Aug 13 07:20:47.483252 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 07:20:47.509174 kernel: libata version 3.00 loaded. Aug 13 07:20:47.509188 kernel: PTP clock support registered Aug 13 07:20:47.509196 kernel: ACPI: bus type USB registered Aug 13 07:20:47.484615 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:20:47.576177 kernel: usbcore: registered new interface driver usbfs Aug 13 07:20:47.576194 kernel: usbcore: registered new interface driver hub Aug 13 07:20:47.576202 kernel: usbcore: registered new device driver usb Aug 13 07:20:47.576209 kernel: AVX2 version of gcm_enc/dec engaged. Aug 13 07:20:47.484647 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:20:47.598176 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:20:47.739398 kernel: AES CTR mode by8 optimization enabled Aug 13 07:20:47.739415 kernel: ahci 0000:00:17.0: version 3.0 Aug 13 07:20:47.739510 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Aug 13 07:20:47.739519 kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 7 ports 6 Gbps 0x7f impl SATA mode Aug 13 07:20:47.739588 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Aug 13 07:20:47.739596 kernel: ahci 0000:00:17.0: flags: 64bit ncq sntf clo only pio slum part ems deso sadm sds apst Aug 13 07:20:47.739659 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Aug 13 07:20:47.739722 kernel: igb 0000:03:00.0: added PHC on eth0 Aug 13 07:20:47.739790 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 1 Aug 13 07:20:47.739851 kernel: igb 0000:03:00.0: Intel(R) Gigabit Ethernet Network Connection Aug 13 07:20:47.739914 kernel: xhci_hcd 0000:00:14.0: hcc params 0x200077c1 hci version 0x110 quirks 0x0000000000009810 Aug 13 07:20:47.739974 kernel: igb 0000:03:00.0: eth0: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:ef:be Aug 13 07:20:47.740038 kernel: scsi host0: ahci Aug 13 07:20:47.740107 kernel: scsi host1: ahci Aug 13 07:20:47.740168 kernel: scsi host2: ahci Aug 13 07:20:47.740226 kernel: scsi host3: ahci Aug 13 07:20:47.740287 kernel: scsi host4: ahci Aug 13 07:20:47.740343 kernel: scsi host5: ahci Aug 13 07:20:47.740402 kernel: scsi host6: ahci Aug 13 07:20:47.740461 kernel: ata1: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516100 irq 127 Aug 13 07:20:47.740469 kernel: ata2: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516180 irq 127 Aug 13 07:20:47.740477 kernel: ata3: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516200 irq 127 Aug 13 07:20:47.740484 kernel: ata4: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516280 irq 127 Aug 13 07:20:47.740491 kernel: ata5: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516300 irq 127 Aug 13 07:20:47.740498 kernel: ata6: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516380 irq 127 Aug 13 07:20:47.740505 kernel: ata7: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516400 irq 127 Aug 13 07:20:47.740514 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Aug 13 07:20:47.740576 kernel: igb 0000:03:00.0: eth0: PBA No: 010000-000 Aug 13 07:20:47.740639 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 2 Aug 13 07:20:47.740699 kernel: igb 0000:03:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Aug 13 07:20:47.740761 kernel: xhci_hcd 0000:00:14.0: Host supports USB 3.1 Enhanced SuperSpeed Aug 13 07:20:47.614153 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:20:48.282095 kernel: igb 0000:04:00.0: added PHC on eth1 Aug 13 07:20:48.282182 kernel: hub 1-0:1.0: USB hub found Aug 13 07:20:48.282258 kernel: igb 0000:04:00.0: Intel(R) Gigabit Ethernet Network Connection Aug 13 07:20:48.282325 kernel: hub 1-0:1.0: 16 ports detected Aug 13 07:20:48.282387 kernel: igb 0000:04:00.0: eth1: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:ef:bf Aug 13 07:20:48.282452 kernel: hub 2-0:1.0: USB hub found Aug 13 07:20:48.282517 kernel: igb 0000:04:00.0: eth1: PBA No: 010000-000 Aug 13 07:20:48.282582 kernel: hub 2-0:1.0: 10 ports detected Aug 13 07:20:48.282643 kernel: igb 0000:04:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Aug 13 07:20:48.282706 kernel: ata1: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Aug 13 07:20:48.282715 kernel: ata5: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282722 kernel: ata1.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Aug 13 07:20:48.282729 kernel: ata4: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282736 kernel: ata6: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282743 kernel: usb 1-14: new high-speed USB device number 2 using xhci_hcd Aug 13 07:20:48.282760 kernel: ata3: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282767 kernel: ata7: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282774 kernel: ata1.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Aug 13 07:20:48.282781 kernel: ata2: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Aug 13 07:20:47.614195 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:20:48.357167 kernel: ata2.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Aug 13 07:20:48.357179 kernel: mlx5_core 0000:01:00.0: firmware version: 14.31.1014 Aug 13 07:20:48.357266 kernel: ata2.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Aug 13 07:20:48.357278 kernel: mlx5_core 0000:01:00.0: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Aug 13 07:20:48.357346 kernel: ata1.00: Features: NCQ-prio Aug 13 07:20:48.205165 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:20:48.437302 kernel: hub 1-14:1.0: USB hub found Aug 13 07:20:48.437391 kernel: ata2.00: Features: NCQ-prio Aug 13 07:20:48.437403 kernel: hub 1-14:1.0: 4 ports detected Aug 13 07:20:48.437471 kernel: ata1.00: configured for UDMA/133 Aug 13 07:20:48.437480 kernel: scsi 0:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Aug 13 07:20:48.387227 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:20:48.465652 kernel: ata2.00: configured for UDMA/133 Aug 13 07:20:48.465663 kernel: scsi 1:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Aug 13 07:20:48.437473 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 07:20:48.466151 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:20:48.483470 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:20:48.483513 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:20:48.500104 kernel: igb 0000:03:00.0 eno1: renamed from eth0 Aug 13 07:20:48.501144 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 07:20:48.984193 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:48.984211 kernel: igb 0000:04:00.0 eno2: renamed from eth1 Aug 13 07:20:48.984307 kernel: ata2.00: Enabling discard_zeroes_data Aug 13 07:20:48.984320 kernel: sd 0:0:0:0: [sdb] 937703088 512-byte logical blocks: (480 GB/447 GiB) Aug 13 07:20:48.984395 kernel: sd 1:0:0:0: [sda] 937703088 512-byte logical blocks: (480 GB/447 GiB) Aug 13 07:20:48.984463 kernel: sd 0:0:0:0: [sdb] 4096-byte physical blocks Aug 13 07:20:48.984525 kernel: sd 0:0:0:0: [sdb] Write Protect is off Aug 13 07:20:48.984587 kernel: sd 1:0:0:0: [sda] 4096-byte physical blocks Aug 13 07:20:48.984647 kernel: sd 0:0:0:0: [sdb] Mode Sense: 00 3a 00 00 Aug 13 07:20:48.984709 kernel: sd 0:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Aug 13 07:20:48.984772 kernel: sd 0:0:0:0: [sdb] Preferred minimum I/O size 4096 bytes Aug 13 07:20:48.984834 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:48.984843 kernel: mlx5_core 0000:01:00.0: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Aug 13 07:20:48.984912 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 07:20:48.984921 kernel: sd 1:0:0:0: [sda] Write Protect is off Aug 13 07:20:48.984982 kernel: mlx5_core 0000:01:00.0: Port module event: module 0, Cable plugged Aug 13 07:20:48.985048 kernel: GPT:9289727 != 937703087 Aug 13 07:20:48.985056 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 07:20:48.985071 kernel: GPT:9289727 != 937703087 Aug 13 07:20:48.985079 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 07:20:48.985086 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:48.985093 kernel: sd 1:0:0:0: [sda] Mode Sense: 00 3a 00 00 Aug 13 07:20:48.985157 kernel: sd 0:0:0:0: [sdb] Attached SCSI disk Aug 13 07:20:48.985219 kernel: sd 1:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Aug 13 07:20:48.985279 kernel: BTRFS: device fsid 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad devid 1 transid 37 /dev/sdb3 scanned by (udev-worker) (558) Aug 13 07:20:48.985288 kernel: sd 1:0:0:0: [sda] Preferred minimum I/O size 4096 bytes Aug 13 07:20:48.985350 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sdb6 scanned by (udev-worker) (541) Aug 13 07:20:48.985358 kernel: usb 1-14.1: new low-speed USB device number 3 using xhci_hcd Aug 13 07:20:48.985464 kernel: ata2.00: Enabling discard_zeroes_data Aug 13 07:20:48.985473 kernel: sd 1:0:0:0: [sda] Attached SCSI disk Aug 13 07:20:48.985536 kernel: hid: raw HID events driver (C) Jiri Kosina Aug 13 07:20:48.958725 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:20:49.003407 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Micron_5300_MTFDDAK480TDT EFI-SYSTEM. Aug 13 07:20:49.074156 kernel: usbcore: registered new interface driver usbhid Aug 13 07:20:49.074169 kernel: usbhid: USB HID core driver Aug 13 07:20:49.074177 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.0/0003:0557:2419.0001/input/input0 Aug 13 07:20:49.054716 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Micron_5300_MTFDDAK480TDT ROOT. Aug 13 07:20:49.088235 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:20:49.281519 kernel: hid-generic 0003:0557:2419.0001: input,hidraw0: USB HID v1.00 Keyboard [HID 0557:2419] on usb-0000:00:14.0-14.1/input0 Aug 13 07:20:49.281653 kernel: mlx5_core 0000:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 07:20:49.281768 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.1/0003:0557:2419.0002/input/input1 Aug 13 07:20:49.281784 kernel: mlx5_core 0000:01:00.1: firmware version: 14.31.1014 Aug 13 07:20:49.281892 kernel: hid-generic 0003:0557:2419.0002: input,hidraw1: USB HID v1.00 Mouse [HID 0557:2419] on usb-0000:00:14.0-14.1/input1 Aug 13 07:20:49.282006 kernel: mlx5_core 0000:01:00.1: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Aug 13 07:20:49.113671 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Micron_5300_MTFDDAK480TDT USR-A. Aug 13 07:20:49.292158 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Micron_5300_MTFDDAK480TDT USR-A. Aug 13 07:20:49.320793 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Aug 13 07:20:49.363227 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 07:20:49.401188 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:49.401202 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:49.379556 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:20:49.425738 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:49.425794 disk-uuid[703]: Primary Header is updated. Aug 13 07:20:49.425794 disk-uuid[703]: Secondary Entries is updated. Aug 13 07:20:49.425794 disk-uuid[703]: Secondary Header is updated. Aug 13 07:20:49.468744 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:49.468755 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:49.489104 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:49.501460 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:20:49.600127 kernel: mlx5_core 0000:01:00.1: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Aug 13 07:20:49.624977 kernel: mlx5_core 0000:01:00.1: Port module event: module 1, Cable plugged Aug 13 07:20:49.927092 kernel: mlx5_core 0000:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 07:20:49.952129 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: renamed from eth0 Aug 13 07:20:49.978144 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: renamed from eth1 Aug 13 07:20:50.467646 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:50.488131 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:50.488179 disk-uuid[704]: The operation has completed successfully. Aug 13 07:20:50.524178 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 07:20:50.524244 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 07:20:50.558354 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 07:20:50.584236 sh[742]: Success Aug 13 07:20:50.594163 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Aug 13 07:20:50.633109 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 07:20:50.643023 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 07:20:50.661059 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 07:20:50.723157 kernel: BTRFS info (device dm-0): first mount of filesystem 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad Aug 13 07:20:50.723203 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:20:50.723236 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Aug 13 07:20:50.723258 kernel: BTRFS info (device dm-0): disabling log replay at mount time Aug 13 07:20:50.723278 kernel: BTRFS info (device dm-0): using free space tree Aug 13 07:20:50.723293 kernel: BTRFS info (device dm-0): enabling ssd optimizations Aug 13 07:20:50.796417 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 07:20:50.805486 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 07:20:50.812349 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 07:20:50.851573 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 07:20:50.893132 kernel: BTRFS info (device sdb6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:20:50.893170 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:20:50.906264 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:20:50.994887 kernel: BTRFS info (device sdb6): using free space tree Aug 13 07:20:50.994902 kernel: BTRFS info (device sdb6): enabling ssd optimizations Aug 13 07:20:50.994909 kernel: BTRFS info (device sdb6): auto enabling async discard Aug 13 07:20:50.994916 kernel: BTRFS info (device sdb6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:20:51.011425 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:20:51.011557 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 07:20:51.022560 systemd-networkd[922]: lo: Link UP Aug 13 07:20:51.022563 systemd-networkd[922]: lo: Gained carrier Aug 13 07:20:51.025133 systemd-networkd[922]: Enumeration completed Aug 13 07:20:51.026117 systemd-networkd[922]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.038390 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:20:51.054034 systemd-networkd[922]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.056594 systemd[1]: Reached target network.target - Network. Aug 13 07:20:51.084043 systemd-networkd[922]: enp1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.089527 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 07:20:51.156684 unknown[927]: fetched base config from "system" Aug 13 07:20:51.153933 ignition[927]: Ignition 2.19.0 Aug 13 07:20:51.156690 unknown[927]: fetched user config from "system" Aug 13 07:20:51.153937 ignition[927]: Stage: fetch-offline Aug 13 07:20:51.157798 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:20:51.153962 ignition[927]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:20:51.174218 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Aug 13 07:20:51.153968 ignition[927]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:20:51.183348 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 07:20:51.154027 ignition[927]: parsed url from cmdline: "" Aug 13 07:20:51.154029 ignition[927]: no config URL provided Aug 13 07:20:51.289249 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Aug 13 07:20:51.284438 systemd-networkd[922]: enp1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.154031 ignition[927]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 07:20:51.154055 ignition[927]: parsing config with SHA512: dbf51eb6cc0653fb3cc72ba56cd8c158fea4eda5e9d0bbcff0ccf5e44f0db0863474bfdbd18a4f57d5872bf9d2e499f2e0a83b1a3b2a27ebec96c34551325a17 Aug 13 07:20:51.156934 ignition[927]: fetch-offline: fetch-offline passed Aug 13 07:20:51.156937 ignition[927]: POST message to Packet Timeline Aug 13 07:20:51.156940 ignition[927]: POST Status error: resource requires networking Aug 13 07:20:51.156977 ignition[927]: Ignition finished successfully Aug 13 07:20:51.191331 ignition[942]: Ignition 2.19.0 Aug 13 07:20:51.191336 ignition[942]: Stage: kargs Aug 13 07:20:51.191463 ignition[942]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:20:51.191471 ignition[942]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:20:51.192172 ignition[942]: kargs: kargs passed Aug 13 07:20:51.192175 ignition[942]: POST message to Packet Timeline Aug 13 07:20:51.192187 ignition[942]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:20:51.192711 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:54753->[::1]:53: read: connection refused Aug 13 07:20:51.393534 ignition[942]: GET https://metadata.packet.net/metadata: attempt #2 Aug 13 07:20:51.394668 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:38785->[::1]:53: read: connection refused Aug 13 07:20:51.547201 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Aug 13 07:20:51.547940 systemd-networkd[922]: eno1: Link UP Aug 13 07:20:51.548116 systemd-networkd[922]: eno2: Link UP Aug 13 07:20:51.548288 systemd-networkd[922]: enp1s0f0np0: Link UP Aug 13 07:20:51.548468 systemd-networkd[922]: enp1s0f0np0: Gained carrier Aug 13 07:20:51.557332 systemd-networkd[922]: enp1s0f1np1: Link UP Aug 13 07:20:51.585235 systemd-networkd[922]: enp1s0f0np0: DHCPv4 address 145.40.90.243/31, gateway 145.40.90.242 acquired from 145.40.83.140 Aug 13 07:20:51.795119 ignition[942]: GET https://metadata.packet.net/metadata: attempt #3 Aug 13 07:20:51.796127 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:50675->[::1]:53: read: connection refused Aug 13 07:20:52.330728 systemd-networkd[922]: enp1s0f1np1: Gained carrier Aug 13 07:20:52.596664 ignition[942]: GET https://metadata.packet.net/metadata: attempt #4 Aug 13 07:20:52.597721 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:38445->[::1]:53: read: connection refused Aug 13 07:20:52.842692 systemd-networkd[922]: enp1s0f0np0: Gained IPv6LL Aug 13 07:20:53.930703 systemd-networkd[922]: enp1s0f1np1: Gained IPv6LL Aug 13 07:20:54.199342 ignition[942]: GET https://metadata.packet.net/metadata: attempt #5 Aug 13 07:20:54.200452 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:55213->[::1]:53: read: connection refused Aug 13 07:20:57.403964 ignition[942]: GET https://metadata.packet.net/metadata: attempt #6 Aug 13 07:20:59.003266 ignition[942]: GET result: OK Aug 13 07:21:00.060590 ignition[942]: Ignition finished successfully Aug 13 07:21:00.065609 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 07:21:00.097315 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 07:21:00.103579 ignition[962]: Ignition 2.19.0 Aug 13 07:21:00.103584 ignition[962]: Stage: disks Aug 13 07:21:00.103700 ignition[962]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:00.103707 ignition[962]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:00.104305 ignition[962]: disks: disks passed Aug 13 07:21:00.104308 ignition[962]: POST message to Packet Timeline Aug 13 07:21:00.104319 ignition[962]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:01.525110 ignition[962]: GET result: OK Aug 13 07:21:02.415724 ignition[962]: Ignition finished successfully Aug 13 07:21:02.418694 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 07:21:02.434341 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 07:21:02.452346 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 07:21:02.473363 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:21:02.495373 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:21:02.516382 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:21:02.550342 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 07:21:02.585397 systemd-fsck[980]: ROOT: clean, 14/553520 files, 52654/553472 blocks Aug 13 07:21:02.595517 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 07:21:02.626484 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 07:21:02.724755 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 07:21:02.739304 kernel: EXT4-fs (sdb9): mounted filesystem 98cc0201-e9ec-4d2c-8a62-5b521bf9317d r/w with ordered data mode. Quota mode: none. Aug 13 07:21:02.725018 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 07:21:02.757273 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:21:02.760857 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 07:21:02.776160 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Aug 13 07:21:02.806069 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sdb6 scanned by mount (989) Aug 13 07:21:02.806085 kernel: BTRFS info (device sdb6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:21:02.838011 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:21:02.842552 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Aug 13 07:21:02.907049 kernel: BTRFS info (device sdb6): using free space tree Aug 13 07:21:02.907071 kernel: BTRFS info (device sdb6): enabling ssd optimizations Aug 13 07:21:02.907085 kernel: BTRFS info (device sdb6): auto enabling async discard Aug 13 07:21:02.907045 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 07:21:02.907072 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:21:02.955619 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:21:02.979327 coreos-metadata[991]: Aug 13 07:21:02.966 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 07:21:02.964355 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 07:21:03.002318 coreos-metadata[992]: Aug 13 07:21:02.966 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 07:21:03.001255 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 07:21:03.040243 initrd-setup-root[1021]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 07:21:03.050179 initrd-setup-root[1028]: cut: /sysroot/etc/group: No such file or directory Aug 13 07:21:03.060181 initrd-setup-root[1035]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 07:21:03.071193 initrd-setup-root[1042]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 07:21:03.087033 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 07:21:03.111339 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 07:21:03.147192 kernel: BTRFS info (device sdb6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:21:03.129806 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 07:21:03.155844 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 07:21:03.177838 ignition[1109]: INFO : Ignition 2.19.0 Aug 13 07:21:03.177838 ignition[1109]: INFO : Stage: mount Aug 13 07:21:03.186178 ignition[1109]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:03.186178 ignition[1109]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:03.186178 ignition[1109]: INFO : mount: mount passed Aug 13 07:21:03.186178 ignition[1109]: INFO : POST message to Packet Timeline Aug 13 07:21:03.186178 ignition[1109]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:03.182829 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 07:21:04.005176 coreos-metadata[991]: Aug 13 07:21:04.005 INFO Fetch successful Aug 13 07:21:04.015161 coreos-metadata[992]: Aug 13 07:21:04.009 INFO Fetch successful Aug 13 07:21:04.036520 coreos-metadata[991]: Aug 13 07:21:04.036 INFO wrote hostname ci-4081.3.5-a-3ca3f1276f to /sysroot/etc/hostname Aug 13 07:21:04.049448 ignition[1109]: INFO : GET result: OK Aug 13 07:21:04.037796 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 07:21:04.060997 systemd[1]: flatcar-static-network.service: Deactivated successfully. Aug 13 07:21:04.061270 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Aug 13 07:21:04.502204 ignition[1109]: INFO : Ignition finished successfully Aug 13 07:21:04.506052 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 07:21:04.533295 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 07:21:04.544108 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:21:04.614827 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sdb6 scanned by mount (1136) Aug 13 07:21:04.614854 kernel: BTRFS info (device sdb6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:21:04.634927 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:21:04.652609 kernel: BTRFS info (device sdb6): using free space tree Aug 13 07:21:04.689968 kernel: BTRFS info (device sdb6): enabling ssd optimizations Aug 13 07:21:04.689984 kernel: BTRFS info (device sdb6): auto enabling async discard Aug 13 07:21:04.702677 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:21:04.730821 ignition[1153]: INFO : Ignition 2.19.0 Aug 13 07:21:04.730821 ignition[1153]: INFO : Stage: files Aug 13 07:21:04.745309 ignition[1153]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:04.745309 ignition[1153]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:04.745309 ignition[1153]: DEBUG : files: compiled without relabeling support, skipping Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 07:21:04.745309 ignition[1153]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Aug 13 07:21:04.735015 unknown[1153]: wrote ssh authorized keys file for user: core Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:05.163421 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Aug 13 07:21:05.644359 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Aug 13 07:21:05.845251 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:05.845251 ignition[1153]: INFO : files: op(c): [started] processing unit "containerd.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(c): [finished] processing unit "containerd.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: files passed Aug 13 07:21:05.874367 ignition[1153]: INFO : POST message to Packet Timeline Aug 13 07:21:05.874367 ignition[1153]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:06.727274 ignition[1153]: INFO : GET result: OK Aug 13 07:21:07.145826 ignition[1153]: INFO : Ignition finished successfully Aug 13 07:21:07.148676 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 07:21:07.187384 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 07:21:07.198801 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 07:21:07.220602 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 07:21:07.220708 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 07:21:07.260907 initrd-setup-root-after-ignition[1191]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:21:07.260907 initrd-setup-root-after-ignition[1191]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:21:07.299380 initrd-setup-root-after-ignition[1195]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:21:07.265371 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:21:07.276341 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 07:21:07.326463 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 07:21:07.410248 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 07:21:07.410300 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 07:21:07.428449 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 07:21:07.439408 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 07:21:07.466440 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 07:21:07.487289 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 07:21:07.530935 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:21:07.561623 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 07:21:07.580658 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:21:07.584261 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:21:07.616496 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 07:21:07.635420 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 07:21:07.635593 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:21:07.663810 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 07:21:07.684704 systemd[1]: Stopped target basic.target - Basic System. Aug 13 07:21:07.703694 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 07:21:07.722794 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:21:07.743680 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 07:21:07.764672 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 07:21:07.785427 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:21:07.806706 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 07:21:07.827773 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 07:21:07.847650 systemd[1]: Stopped target swap.target - Swaps. Aug 13 07:21:07.865662 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 07:21:07.866088 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:21:07.900556 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:21:07.910703 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:21:07.932549 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 07:21:07.932910 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:21:07.954570 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 07:21:07.954975 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 07:21:07.987662 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 07:21:07.988154 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:21:08.008890 systemd[1]: Stopped target paths.target - Path Units. Aug 13 07:21:08.027545 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 07:21:08.027969 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:21:08.048700 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 07:21:08.066653 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 07:21:08.084665 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 07:21:08.084973 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:21:08.104738 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 07:21:08.105044 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:21:08.127777 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 07:21:08.128210 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:21:08.148770 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 07:21:08.265289 ignition[1215]: INFO : Ignition 2.19.0 Aug 13 07:21:08.265289 ignition[1215]: INFO : Stage: umount Aug 13 07:21:08.265289 ignition[1215]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:08.265289 ignition[1215]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:08.265289 ignition[1215]: INFO : umount: umount passed Aug 13 07:21:08.265289 ignition[1215]: INFO : POST message to Packet Timeline Aug 13 07:21:08.265289 ignition[1215]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:08.149181 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 07:21:08.166783 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Aug 13 07:21:08.167195 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 07:21:08.196241 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 07:21:08.211360 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 07:21:08.211489 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:21:08.249268 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 07:21:08.250406 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 07:21:08.250488 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:21:08.276331 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 07:21:08.276411 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:21:08.326344 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 07:21:08.328002 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 07:21:08.328279 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 07:21:08.341940 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 07:21:08.342201 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 07:21:09.150216 ignition[1215]: INFO : GET result: OK Aug 13 07:21:09.595137 ignition[1215]: INFO : Ignition finished successfully Aug 13 07:21:09.598209 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 07:21:09.598515 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 07:21:09.615396 systemd[1]: Stopped target network.target - Network. Aug 13 07:21:09.631303 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 07:21:09.631486 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 07:21:09.649488 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 07:21:09.649654 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 07:21:09.669402 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 07:21:09.669544 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 07:21:09.688475 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 07:21:09.688647 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 07:21:09.707471 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 07:21:09.707643 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 07:21:09.726878 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 07:21:09.737203 systemd-networkd[922]: enp1s0f1np1: DHCPv6 lease lost Aug 13 07:21:09.744296 systemd-networkd[922]: enp1s0f0np0: DHCPv6 lease lost Aug 13 07:21:09.744585 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 07:21:09.765210 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 07:21:09.765499 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 07:21:09.785444 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 07:21:09.785790 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 07:21:09.806191 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 07:21:09.806378 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:21:09.837319 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 07:21:09.844441 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 07:21:09.844599 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:21:09.872412 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 07:21:09.872560 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:21:09.890362 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 07:21:09.890503 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 07:21:09.908471 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 07:21:09.908640 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:21:09.929799 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:21:09.966846 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 07:21:09.967271 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:21:09.971672 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 07:21:09.971829 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 07:21:09.999423 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 07:21:09.999537 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:21:10.017335 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 07:21:10.017479 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:21:10.047627 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 07:21:10.047798 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 07:21:10.078496 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:21:10.078648 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:21:10.139261 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 07:21:10.168225 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 07:21:10.168270 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:21:10.186224 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:21:10.422260 systemd-journald[267]: Received SIGTERM from PID 1 (systemd). Aug 13 07:21:10.186281 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:21:10.210455 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 07:21:10.210712 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 07:21:10.271623 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 07:21:10.271922 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 07:21:10.291085 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 07:21:10.330458 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 07:21:10.350722 systemd[1]: Switching root. Aug 13 07:21:10.495286 systemd-journald[267]: Journal stopped Aug 13 07:20:46.018996 kernel: Linux version 6.6.100-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Aug 12 22:14:58 -00 2025 Aug 13 07:20:46.019011 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:20:46.019019 kernel: BIOS-provided physical RAM map: Aug 13 07:20:46.019023 kernel: BIOS-e820: [mem 0x0000000000000000-0x00000000000997ff] usable Aug 13 07:20:46.019027 kernel: BIOS-e820: [mem 0x0000000000099800-0x000000000009ffff] reserved Aug 13 07:20:46.019031 kernel: BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved Aug 13 07:20:46.019035 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003fffffff] usable Aug 13 07:20:46.019040 kernel: BIOS-e820: [mem 0x0000000040000000-0x00000000403fffff] reserved Aug 13 07:20:46.019044 kernel: BIOS-e820: [mem 0x0000000040400000-0x00000000819c6fff] usable Aug 13 07:20:46.019048 kernel: BIOS-e820: [mem 0x00000000819c7000-0x00000000819c7fff] ACPI NVS Aug 13 07:20:46.019052 kernel: BIOS-e820: [mem 0x00000000819c8000-0x00000000819c8fff] reserved Aug 13 07:20:46.019057 kernel: BIOS-e820: [mem 0x00000000819c9000-0x000000008afcdfff] usable Aug 13 07:20:46.019061 kernel: BIOS-e820: [mem 0x000000008afce000-0x000000008c0b2fff] reserved Aug 13 07:20:46.019068 kernel: BIOS-e820: [mem 0x000000008c0b3000-0x000000008c23bfff] usable Aug 13 07:20:46.019073 kernel: BIOS-e820: [mem 0x000000008c23c000-0x000000008c66dfff] ACPI NVS Aug 13 07:20:46.019078 kernel: BIOS-e820: [mem 0x000000008c66e000-0x000000008eefefff] reserved Aug 13 07:20:46.019102 kernel: BIOS-e820: [mem 0x000000008eeff000-0x000000008eefffff] usable Aug 13 07:20:46.019107 kernel: BIOS-e820: [mem 0x000000008ef00000-0x000000008fffffff] reserved Aug 13 07:20:46.019112 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Aug 13 07:20:46.019117 kernel: BIOS-e820: [mem 0x00000000fe000000-0x00000000fe010fff] reserved Aug 13 07:20:46.019135 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec00fff] reserved Aug 13 07:20:46.019139 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Aug 13 07:20:46.019144 kernel: BIOS-e820: [mem 0x00000000ff000000-0x00000000ffffffff] reserved Aug 13 07:20:46.019148 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000086effffff] usable Aug 13 07:20:46.019153 kernel: NX (Execute Disable) protection: active Aug 13 07:20:46.019158 kernel: APIC: Static calls initialized Aug 13 07:20:46.019162 kernel: SMBIOS 3.2.1 present. Aug 13 07:20:46.019167 kernel: DMI: Supermicro SYS-5019C-MR-PH004/X11SCM-F, BIOS 2.6 12/03/2024 Aug 13 07:20:46.019173 kernel: tsc: Detected 3400.000 MHz processor Aug 13 07:20:46.019177 kernel: tsc: Detected 3399.906 MHz TSC Aug 13 07:20:46.019182 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Aug 13 07:20:46.019187 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Aug 13 07:20:46.019192 kernel: last_pfn = 0x86f000 max_arch_pfn = 0x400000000 Aug 13 07:20:46.019197 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 23), built from 10 variable MTRRs Aug 13 07:20:46.019201 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Aug 13 07:20:46.019206 kernel: last_pfn = 0x8ef00 max_arch_pfn = 0x400000000 Aug 13 07:20:46.019211 kernel: Using GB pages for direct mapping Aug 13 07:20:46.019217 kernel: ACPI: Early table checksum verification disabled Aug 13 07:20:46.019221 kernel: ACPI: RSDP 0x00000000000F05B0 000024 (v02 SUPERM) Aug 13 07:20:46.019226 kernel: ACPI: XSDT 0x000000008C54F0C8 00010C (v01 SUPERM SUPERM 01072009 AMI 00010013) Aug 13 07:20:46.019233 kernel: ACPI: FACP 0x000000008C58B670 000114 (v06 01072009 AMI 00010013) Aug 13 07:20:46.019238 kernel: ACPI: DSDT 0x000000008C54F268 03C404 (v02 SUPERM SMCI--MB 01072009 INTL 20160527) Aug 13 07:20:46.019243 kernel: ACPI: FACS 0x000000008C66DF80 000040 Aug 13 07:20:46.019248 kernel: ACPI: APIC 0x000000008C58B788 00012C (v04 01072009 AMI 00010013) Aug 13 07:20:46.019254 kernel: ACPI: FPDT 0x000000008C58B8B8 000044 (v01 01072009 AMI 00010013) Aug 13 07:20:46.019259 kernel: ACPI: FIDT 0x000000008C58B900 00009C (v01 SUPERM SMCI--MB 01072009 AMI 00010013) Aug 13 07:20:46.019264 kernel: ACPI: MCFG 0x000000008C58B9A0 00003C (v01 SUPERM SMCI--MB 01072009 MSFT 00000097) Aug 13 07:20:46.019269 kernel: ACPI: SPMI 0x000000008C58B9E0 000041 (v05 SUPERM SMCI--MB 00000000 AMI. 00000000) Aug 13 07:20:46.019274 kernel: ACPI: SSDT 0x000000008C58BA28 001B1C (v02 CpuRef CpuSsdt 00003000 INTL 20160527) Aug 13 07:20:46.019279 kernel: ACPI: SSDT 0x000000008C58D548 0031C6 (v02 SaSsdt SaSsdt 00003000 INTL 20160527) Aug 13 07:20:46.019283 kernel: ACPI: SSDT 0x000000008C590710 00232B (v02 PegSsd PegSsdt 00001000 INTL 20160527) Aug 13 07:20:46.019289 kernel: ACPI: HPET 0x000000008C592A40 000038 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019294 kernel: ACPI: SSDT 0x000000008C592A78 000FAE (v02 SUPERM Ther_Rvp 00001000 INTL 20160527) Aug 13 07:20:46.019299 kernel: ACPI: SSDT 0x000000008C593A28 0008F4 (v02 INTEL xh_mossb 00000000 INTL 20160527) Aug 13 07:20:46.019304 kernel: ACPI: UEFI 0x000000008C594320 000042 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019309 kernel: ACPI: LPIT 0x000000008C594368 000094 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019314 kernel: ACPI: SSDT 0x000000008C594400 0027DE (v02 SUPERM PtidDevc 00001000 INTL 20160527) Aug 13 07:20:46.019319 kernel: ACPI: SSDT 0x000000008C596BE0 0014E2 (v02 SUPERM TbtTypeC 00000000 INTL 20160527) Aug 13 07:20:46.019324 kernel: ACPI: DBGP 0x000000008C5980C8 000034 (v01 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019329 kernel: ACPI: DBG2 0x000000008C598100 000054 (v00 SUPERM SMCI--MB 00000002 01000013) Aug 13 07:20:46.019335 kernel: ACPI: SSDT 0x000000008C598158 001B67 (v02 SUPERM UsbCTabl 00001000 INTL 20160527) Aug 13 07:20:46.019340 kernel: ACPI: DMAR 0x000000008C599CC0 000070 (v01 INTEL EDK2 00000002 01000013) Aug 13 07:20:46.019345 kernel: ACPI: SSDT 0x000000008C599D30 000144 (v02 Intel ADebTabl 00001000 INTL 20160527) Aug 13 07:20:46.019349 kernel: ACPI: TPM2 0x000000008C599E78 000034 (v04 SUPERM SMCI--MB 00000001 AMI 00000000) Aug 13 07:20:46.019354 kernel: ACPI: SSDT 0x000000008C599EB0 000D8F (v02 INTEL SpsNm 00000002 INTL 20160527) Aug 13 07:20:46.019359 kernel: ACPI: WSMT 0x000000008C59AC40 000028 (v01 SUPERM 01072009 AMI 00010013) Aug 13 07:20:46.019364 kernel: ACPI: EINJ 0x000000008C59AC68 000130 (v01 AMI AMI.EINJ 00000000 AMI. 00000000) Aug 13 07:20:46.019369 kernel: ACPI: ERST 0x000000008C59AD98 000230 (v01 AMIER AMI.ERST 00000000 AMI. 00000000) Aug 13 07:20:46.019375 kernel: ACPI: BERT 0x000000008C59AFC8 000030 (v01 AMI AMI.BERT 00000000 AMI. 00000000) Aug 13 07:20:46.019380 kernel: ACPI: HEST 0x000000008C59AFF8 00027C (v01 AMI AMI.HEST 00000000 AMI. 00000000) Aug 13 07:20:46.019385 kernel: ACPI: SSDT 0x000000008C59B278 000162 (v01 SUPERM SMCCDN 00000000 INTL 20181221) Aug 13 07:20:46.019390 kernel: ACPI: Reserving FACP table memory at [mem 0x8c58b670-0x8c58b783] Aug 13 07:20:46.019395 kernel: ACPI: Reserving DSDT table memory at [mem 0x8c54f268-0x8c58b66b] Aug 13 07:20:46.019400 kernel: ACPI: Reserving FACS table memory at [mem 0x8c66df80-0x8c66dfbf] Aug 13 07:20:46.019405 kernel: ACPI: Reserving APIC table memory at [mem 0x8c58b788-0x8c58b8b3] Aug 13 07:20:46.019410 kernel: ACPI: Reserving FPDT table memory at [mem 0x8c58b8b8-0x8c58b8fb] Aug 13 07:20:46.019415 kernel: ACPI: Reserving FIDT table memory at [mem 0x8c58b900-0x8c58b99b] Aug 13 07:20:46.019420 kernel: ACPI: Reserving MCFG table memory at [mem 0x8c58b9a0-0x8c58b9db] Aug 13 07:20:46.019425 kernel: ACPI: Reserving SPMI table memory at [mem 0x8c58b9e0-0x8c58ba20] Aug 13 07:20:46.019430 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58ba28-0x8c58d543] Aug 13 07:20:46.019435 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58d548-0x8c59070d] Aug 13 07:20:46.019440 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c590710-0x8c592a3a] Aug 13 07:20:46.019445 kernel: ACPI: Reserving HPET table memory at [mem 0x8c592a40-0x8c592a77] Aug 13 07:20:46.019450 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c592a78-0x8c593a25] Aug 13 07:20:46.019455 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c593a28-0x8c59431b] Aug 13 07:20:46.019459 kernel: ACPI: Reserving UEFI table memory at [mem 0x8c594320-0x8c594361] Aug 13 07:20:46.019465 kernel: ACPI: Reserving LPIT table memory at [mem 0x8c594368-0x8c5943fb] Aug 13 07:20:46.019470 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c594400-0x8c596bdd] Aug 13 07:20:46.019475 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c596be0-0x8c5980c1] Aug 13 07:20:46.019480 kernel: ACPI: Reserving DBGP table memory at [mem 0x8c5980c8-0x8c5980fb] Aug 13 07:20:46.019485 kernel: ACPI: Reserving DBG2 table memory at [mem 0x8c598100-0x8c598153] Aug 13 07:20:46.019490 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c598158-0x8c599cbe] Aug 13 07:20:46.019494 kernel: ACPI: Reserving DMAR table memory at [mem 0x8c599cc0-0x8c599d2f] Aug 13 07:20:46.019499 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599d30-0x8c599e73] Aug 13 07:20:46.019504 kernel: ACPI: Reserving TPM2 table memory at [mem 0x8c599e78-0x8c599eab] Aug 13 07:20:46.019510 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599eb0-0x8c59ac3e] Aug 13 07:20:46.019515 kernel: ACPI: Reserving WSMT table memory at [mem 0x8c59ac40-0x8c59ac67] Aug 13 07:20:46.019520 kernel: ACPI: Reserving EINJ table memory at [mem 0x8c59ac68-0x8c59ad97] Aug 13 07:20:46.019525 kernel: ACPI: Reserving ERST table memory at [mem 0x8c59ad98-0x8c59afc7] Aug 13 07:20:46.019530 kernel: ACPI: Reserving BERT table memory at [mem 0x8c59afc8-0x8c59aff7] Aug 13 07:20:46.019534 kernel: ACPI: Reserving HEST table memory at [mem 0x8c59aff8-0x8c59b273] Aug 13 07:20:46.019539 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c59b278-0x8c59b3d9] Aug 13 07:20:46.019544 kernel: No NUMA configuration found Aug 13 07:20:46.019549 kernel: Faking a node at [mem 0x0000000000000000-0x000000086effffff] Aug 13 07:20:46.019554 kernel: NODE_DATA(0) allocated [mem 0x86effa000-0x86effffff] Aug 13 07:20:46.019560 kernel: Zone ranges: Aug 13 07:20:46.019565 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Aug 13 07:20:46.019570 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Aug 13 07:20:46.019575 kernel: Normal [mem 0x0000000100000000-0x000000086effffff] Aug 13 07:20:46.019580 kernel: Movable zone start for each node Aug 13 07:20:46.019585 kernel: Early memory node ranges Aug 13 07:20:46.019590 kernel: node 0: [mem 0x0000000000001000-0x0000000000098fff] Aug 13 07:20:46.019595 kernel: node 0: [mem 0x0000000000100000-0x000000003fffffff] Aug 13 07:20:46.019600 kernel: node 0: [mem 0x0000000040400000-0x00000000819c6fff] Aug 13 07:20:46.019606 kernel: node 0: [mem 0x00000000819c9000-0x000000008afcdfff] Aug 13 07:20:46.019611 kernel: node 0: [mem 0x000000008c0b3000-0x000000008c23bfff] Aug 13 07:20:46.019615 kernel: node 0: [mem 0x000000008eeff000-0x000000008eefffff] Aug 13 07:20:46.019621 kernel: node 0: [mem 0x0000000100000000-0x000000086effffff] Aug 13 07:20:46.019630 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000086effffff] Aug 13 07:20:46.019635 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Aug 13 07:20:46.019640 kernel: On node 0, zone DMA: 103 pages in unavailable ranges Aug 13 07:20:46.019645 kernel: On node 0, zone DMA32: 1024 pages in unavailable ranges Aug 13 07:20:46.019652 kernel: On node 0, zone DMA32: 2 pages in unavailable ranges Aug 13 07:20:46.019657 kernel: On node 0, zone DMA32: 4325 pages in unavailable ranges Aug 13 07:20:46.019662 kernel: On node 0, zone DMA32: 11459 pages in unavailable ranges Aug 13 07:20:46.019667 kernel: On node 0, zone Normal: 4352 pages in unavailable ranges Aug 13 07:20:46.019673 kernel: On node 0, zone Normal: 4096 pages in unavailable ranges Aug 13 07:20:46.019678 kernel: ACPI: PM-Timer IO Port: 0x1808 Aug 13 07:20:46.019683 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Aug 13 07:20:46.019689 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Aug 13 07:20:46.019694 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Aug 13 07:20:46.019700 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Aug 13 07:20:46.019705 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Aug 13 07:20:46.019710 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Aug 13 07:20:46.019716 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Aug 13 07:20:46.019721 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Aug 13 07:20:46.019726 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Aug 13 07:20:46.019731 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Aug 13 07:20:46.019736 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Aug 13 07:20:46.019741 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Aug 13 07:20:46.019748 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Aug 13 07:20:46.019753 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Aug 13 07:20:46.019758 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Aug 13 07:20:46.019763 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Aug 13 07:20:46.019768 kernel: IOAPIC[0]: apic_id 2, version 32, address 0xfec00000, GSI 0-119 Aug 13 07:20:46.019774 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Aug 13 07:20:46.019779 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Aug 13 07:20:46.019784 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Aug 13 07:20:46.019789 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Aug 13 07:20:46.019796 kernel: TSC deadline timer available Aug 13 07:20:46.019801 kernel: smpboot: Allowing 16 CPUs, 0 hotplug CPUs Aug 13 07:20:46.019806 kernel: [mem 0x90000000-0xdfffffff] available for PCI devices Aug 13 07:20:46.019812 kernel: Booting paravirtualized kernel on bare hardware Aug 13 07:20:46.019817 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Aug 13 07:20:46.019823 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Aug 13 07:20:46.019828 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u262144 Aug 13 07:20:46.019833 kernel: pcpu-alloc: s197096 r8192 d32280 u262144 alloc=1*2097152 Aug 13 07:20:46.019838 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Aug 13 07:20:46.019845 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:20:46.019851 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 07:20:46.019856 kernel: random: crng init done Aug 13 07:20:46.019861 kernel: Dentry cache hash table entries: 4194304 (order: 13, 33554432 bytes, linear) Aug 13 07:20:46.019866 kernel: Inode-cache hash table entries: 2097152 (order: 12, 16777216 bytes, linear) Aug 13 07:20:46.019872 kernel: Fallback order for Node 0: 0 Aug 13 07:20:46.019877 kernel: Built 1 zonelists, mobility grouping on. Total pages: 8232416 Aug 13 07:20:46.019882 kernel: Policy zone: Normal Aug 13 07:20:46.019888 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 07:20:46.019894 kernel: software IO TLB: area num 16. Aug 13 07:20:46.019899 kernel: Memory: 32720316K/33452984K available (12288K kernel code, 2295K rwdata, 22748K rodata, 42876K init, 2316K bss, 732408K reserved, 0K cma-reserved) Aug 13 07:20:46.019904 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Aug 13 07:20:46.019910 kernel: ftrace: allocating 37968 entries in 149 pages Aug 13 07:20:46.019915 kernel: ftrace: allocated 149 pages with 4 groups Aug 13 07:20:46.019920 kernel: Dynamic Preempt: voluntary Aug 13 07:20:46.019926 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 07:20:46.019931 kernel: rcu: RCU event tracing is enabled. Aug 13 07:20:46.019938 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Aug 13 07:20:46.019943 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 07:20:46.019948 kernel: Rude variant of Tasks RCU enabled. Aug 13 07:20:46.019954 kernel: Tracing variant of Tasks RCU enabled. Aug 13 07:20:46.019959 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 07:20:46.019964 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Aug 13 07:20:46.019970 kernel: NR_IRQS: 33024, nr_irqs: 2184, preallocated irqs: 16 Aug 13 07:20:46.019975 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 07:20:46.019980 kernel: Console: colour dummy device 80x25 Aug 13 07:20:46.019985 kernel: printk: console [tty0] enabled Aug 13 07:20:46.019991 kernel: printk: console [ttyS1] enabled Aug 13 07:20:46.019997 kernel: ACPI: Core revision 20230628 Aug 13 07:20:46.020002 kernel: hpet: HPET dysfunctional in PC10. Force disabled. Aug 13 07:20:46.020007 kernel: APIC: Switch to symmetric I/O mode setup Aug 13 07:20:46.020012 kernel: DMAR: Host address width 39 Aug 13 07:20:46.020018 kernel: DMAR: DRHD base: 0x000000fed91000 flags: 0x1 Aug 13 07:20:46.020023 kernel: DMAR: dmar0: reg_base_addr fed91000 ver 1:0 cap d2008c40660462 ecap f050da Aug 13 07:20:46.020029 kernel: DMAR: RMRR base: 0x0000008cf19000 end: 0x0000008d162fff Aug 13 07:20:46.020034 kernel: DMAR-IR: IOAPIC id 2 under DRHD base 0xfed91000 IOMMU 0 Aug 13 07:20:46.020040 kernel: DMAR-IR: HPET id 0 under DRHD base 0xfed91000 Aug 13 07:20:46.020045 kernel: DMAR-IR: Queued invalidation will be enabled to support x2apic and Intr-remapping. Aug 13 07:20:46.020050 kernel: DMAR-IR: Enabled IRQ remapping in x2apic mode Aug 13 07:20:46.020056 kernel: x2apic enabled Aug 13 07:20:46.020061 kernel: APIC: Switched APIC routing to: cluster x2apic Aug 13 07:20:46.020068 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x3101f59f5e6, max_idle_ns: 440795259996 ns Aug 13 07:20:46.020074 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 6799.81 BogoMIPS (lpj=3399906) Aug 13 07:20:46.020098 kernel: CPU0: Thermal monitoring enabled (TM1) Aug 13 07:20:46.020103 kernel: process: using mwait in idle threads Aug 13 07:20:46.020110 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Aug 13 07:20:46.020115 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Aug 13 07:20:46.020134 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Aug 13 07:20:46.020139 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Aug 13 07:20:46.020144 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Aug 13 07:20:46.020149 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Aug 13 07:20:46.020155 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Aug 13 07:20:46.020160 kernel: RETBleed: Mitigation: Enhanced IBRS Aug 13 07:20:46.020165 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Aug 13 07:20:46.020170 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Aug 13 07:20:46.020175 kernel: TAA: Mitigation: TSX disabled Aug 13 07:20:46.020181 kernel: MMIO Stale Data: Mitigation: Clear CPU buffers Aug 13 07:20:46.020187 kernel: SRBDS: Mitigation: Microcode Aug 13 07:20:46.020192 kernel: GDS: Mitigation: Microcode Aug 13 07:20:46.020197 kernel: ITS: Mitigation: Aligned branch/return thunks Aug 13 07:20:46.020202 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Aug 13 07:20:46.020207 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Aug 13 07:20:46.020213 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Aug 13 07:20:46.020218 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Aug 13 07:20:46.020223 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Aug 13 07:20:46.020228 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Aug 13 07:20:46.020233 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Aug 13 07:20:46.020240 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Aug 13 07:20:46.020245 kernel: x86/fpu: Enabled xstate features 0x1f, context size is 960 bytes, using 'compacted' format. Aug 13 07:20:46.020250 kernel: Freeing SMP alternatives memory: 32K Aug 13 07:20:46.020255 kernel: pid_max: default: 32768 minimum: 301 Aug 13 07:20:46.020261 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Aug 13 07:20:46.020266 kernel: landlock: Up and running. Aug 13 07:20:46.020271 kernel: SELinux: Initializing. Aug 13 07:20:46.020276 kernel: Mount-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.020281 kernel: Mountpoint-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.020287 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Aug 13 07:20:46.020292 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Aug 13 07:20:46.020298 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Aug 13 07:20:46.020304 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Aug 13 07:20:46.020309 kernel: Performance Events: PEBS fmt3+, Skylake events, 32-deep LBR, full-width counters, Intel PMU driver. Aug 13 07:20:46.020314 kernel: ... version: 4 Aug 13 07:20:46.020320 kernel: ... bit width: 48 Aug 13 07:20:46.020325 kernel: ... generic registers: 4 Aug 13 07:20:46.020330 kernel: ... value mask: 0000ffffffffffff Aug 13 07:20:46.020335 kernel: ... max period: 00007fffffffffff Aug 13 07:20:46.020341 kernel: ... fixed-purpose events: 3 Aug 13 07:20:46.020347 kernel: ... event mask: 000000070000000f Aug 13 07:20:46.020352 kernel: signal: max sigframe size: 2032 Aug 13 07:20:46.020357 kernel: Estimated ratio of average max frequency by base frequency (times 1024): 1445 Aug 13 07:20:46.020363 kernel: rcu: Hierarchical SRCU implementation. Aug 13 07:20:46.020368 kernel: rcu: Max phase no-delay instances is 400. Aug 13 07:20:46.020373 kernel: NMI watchdog: Enabled. Permanently consumes one hw-PMU counter. Aug 13 07:20:46.020379 kernel: smp: Bringing up secondary CPUs ... Aug 13 07:20:46.020384 kernel: smpboot: x86: Booting SMP configuration: Aug 13 07:20:46.020389 kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6 #7 #8 #9 #10 #11 #12 #13 #14 #15 Aug 13 07:20:46.020396 kernel: Transient Scheduler Attacks: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Aug 13 07:20:46.020401 kernel: smp: Brought up 1 node, 16 CPUs Aug 13 07:20:46.020406 kernel: smpboot: Max logical packages: 1 Aug 13 07:20:46.020412 kernel: smpboot: Total of 16 processors activated (108796.99 BogoMIPS) Aug 13 07:20:46.020417 kernel: devtmpfs: initialized Aug 13 07:20:46.020422 kernel: x86/mm: Memory block size: 128MB Aug 13 07:20:46.020427 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x819c7000-0x819c7fff] (4096 bytes) Aug 13 07:20:46.020433 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x8c23c000-0x8c66dfff] (4399104 bytes) Aug 13 07:20:46.020438 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 07:20:46.020444 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Aug 13 07:20:46.020450 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 07:20:46.020455 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 07:20:46.020460 kernel: audit: initializing netlink subsys (disabled) Aug 13 07:20:46.020465 kernel: audit: type=2000 audit(1755069640.039:1): state=initialized audit_enabled=0 res=1 Aug 13 07:20:46.020470 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 07:20:46.020476 kernel: thermal_sys: Registered thermal governor 'user_space' Aug 13 07:20:46.020481 kernel: cpuidle: using governor menu Aug 13 07:20:46.020486 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 07:20:46.020492 kernel: dca service started, version 1.12.1 Aug 13 07:20:46.020498 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Aug 13 07:20:46.020503 kernel: PCI: Using configuration type 1 for base access Aug 13 07:20:46.020508 kernel: ENERGY_PERF_BIAS: Set to 'normal', was 'performance' Aug 13 07:20:46.020513 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Aug 13 07:20:46.020519 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 07:20:46.020524 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 07:20:46.020529 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 07:20:46.020535 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 07:20:46.020540 kernel: ACPI: Added _OSI(Module Device) Aug 13 07:20:46.020545 kernel: ACPI: Added _OSI(Processor Device) Aug 13 07:20:46.020551 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 07:20:46.020556 kernel: ACPI: 12 ACPI AML tables successfully acquired and loaded Aug 13 07:20:46.020561 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020566 kernel: ACPI: SSDT 0xFFFF9BA581AF5400 000400 (v02 PmRef Cpu0Cst 00003001 INTL 20160527) Aug 13 07:20:46.020572 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020577 kernel: ACPI: SSDT 0xFFFF9BA581AEF800 000683 (v02 PmRef Cpu0Ist 00003000 INTL 20160527) Aug 13 07:20:46.020582 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020588 kernel: ACPI: SSDT 0xFFFF9BA580247E00 0000F4 (v02 PmRef Cpu0Psd 00003000 INTL 20160527) Aug 13 07:20:46.020594 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020599 kernel: ACPI: SSDT 0xFFFF9BA581AEB800 0005FC (v02 PmRef ApIst 00003000 INTL 20160527) Aug 13 07:20:46.020604 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020609 kernel: ACPI: SSDT 0xFFFF9BA58012D000 000AB0 (v02 PmRef ApPsd 00003000 INTL 20160527) Aug 13 07:20:46.020614 kernel: ACPI: Dynamic OEM Table Load: Aug 13 07:20:46.020620 kernel: ACPI: SSDT 0xFFFF9BA581AF0800 00030A (v02 PmRef ApCst 00003000 INTL 20160527) Aug 13 07:20:46.020625 kernel: ACPI: _OSC evaluated successfully for all CPUs Aug 13 07:20:46.020630 kernel: ACPI: Interpreter enabled Aug 13 07:20:46.020636 kernel: ACPI: PM: (supports S0 S5) Aug 13 07:20:46.020642 kernel: ACPI: Using IOAPIC for interrupt routing Aug 13 07:20:46.020647 kernel: HEST: Enabling Firmware First mode for corrected errors. Aug 13 07:20:46.020652 kernel: mce: [Firmware Bug]: Ignoring request to disable invalid MCA bank 14. Aug 13 07:20:46.020657 kernel: HEST: Table parsing has been initialized. Aug 13 07:20:46.020663 kernel: GHES: APEI firmware first mode is enabled by APEI bit and WHEA _OSC. Aug 13 07:20:46.020668 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Aug 13 07:20:46.020673 kernel: PCI: Ignoring E820 reservations for host bridge windows Aug 13 07:20:46.020678 kernel: ACPI: Enabled 9 GPEs in block 00 to 7F Aug 13 07:20:46.020685 kernel: ACPI: \_SB_.PCI0.XDCI.USBC: New power resource Aug 13 07:20:46.020690 kernel: ACPI: \_SB_.PCI0.SAT0.VOL0.V0PR: New power resource Aug 13 07:20:46.020695 kernel: ACPI: \_SB_.PCI0.SAT0.VOL1.V1PR: New power resource Aug 13 07:20:46.020701 kernel: ACPI: \_SB_.PCI0.SAT0.VOL2.V2PR: New power resource Aug 13 07:20:46.020706 kernel: ACPI: \_SB_.PCI0.CNVW.WRST: New power resource Aug 13 07:20:46.020711 kernel: ACPI: \_TZ_.FN00: New power resource Aug 13 07:20:46.020717 kernel: ACPI: \_TZ_.FN01: New power resource Aug 13 07:20:46.020722 kernel: ACPI: \_TZ_.FN02: New power resource Aug 13 07:20:46.020727 kernel: ACPI: \_TZ_.FN03: New power resource Aug 13 07:20:46.020733 kernel: ACPI: \_TZ_.FN04: New power resource Aug 13 07:20:46.020739 kernel: ACPI: \PIN_: New power resource Aug 13 07:20:46.020744 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-fe]) Aug 13 07:20:46.020818 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 07:20:46.020873 kernel: acpi PNP0A08:00: _OSC: platform does not support [AER] Aug 13 07:20:46.020923 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability LTR] Aug 13 07:20:46.020931 kernel: PCI host bridge to bus 0000:00 Aug 13 07:20:46.020983 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Aug 13 07:20:46.021031 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Aug 13 07:20:46.021077 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Aug 13 07:20:46.021158 kernel: pci_bus 0000:00: root bus resource [mem 0x90000000-0xdfffffff window] Aug 13 07:20:46.021200 kernel: pci_bus 0000:00: root bus resource [mem 0xfc800000-0xfe7fffff window] Aug 13 07:20:46.021244 kernel: pci_bus 0000:00: root bus resource [bus 00-fe] Aug 13 07:20:46.021302 kernel: pci 0000:00:00.0: [8086:3e31] type 00 class 0x060000 Aug 13 07:20:46.021361 kernel: pci 0000:00:01.0: [8086:1901] type 01 class 0x060400 Aug 13 07:20:46.021413 kernel: pci 0000:00:01.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.021468 kernel: pci 0000:00:08.0: [8086:1911] type 00 class 0x088000 Aug 13 07:20:46.021518 kernel: pci 0000:00:08.0: reg 0x10: [mem 0x9551f000-0x9551ffff 64bit] Aug 13 07:20:46.021571 kernel: pci 0000:00:12.0: [8086:a379] type 00 class 0x118000 Aug 13 07:20:46.021621 kernel: pci 0000:00:12.0: reg 0x10: [mem 0x9551e000-0x9551efff 64bit] Aug 13 07:20:46.021679 kernel: pci 0000:00:14.0: [8086:a36d] type 00 class 0x0c0330 Aug 13 07:20:46.021729 kernel: pci 0000:00:14.0: reg 0x10: [mem 0x95500000-0x9550ffff 64bit] Aug 13 07:20:46.021778 kernel: pci 0000:00:14.0: PME# supported from D3hot D3cold Aug 13 07:20:46.021832 kernel: pci 0000:00:14.2: [8086:a36f] type 00 class 0x050000 Aug 13 07:20:46.021881 kernel: pci 0000:00:14.2: reg 0x10: [mem 0x95512000-0x95513fff 64bit] Aug 13 07:20:46.021929 kernel: pci 0000:00:14.2: reg 0x18: [mem 0x9551d000-0x9551dfff 64bit] Aug 13 07:20:46.021982 kernel: pci 0000:00:15.0: [8086:a368] type 00 class 0x0c8000 Aug 13 07:20:46.022035 kernel: pci 0000:00:15.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Aug 13 07:20:46.022120 kernel: pci 0000:00:15.1: [8086:a369] type 00 class 0x0c8000 Aug 13 07:20:46.022170 kernel: pci 0000:00:15.1: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Aug 13 07:20:46.022223 kernel: pci 0000:00:16.0: [8086:a360] type 00 class 0x078000 Aug 13 07:20:46.022272 kernel: pci 0000:00:16.0: reg 0x10: [mem 0x9551a000-0x9551afff 64bit] Aug 13 07:20:46.022325 kernel: pci 0000:00:16.0: PME# supported from D3hot Aug 13 07:20:46.022377 kernel: pci 0000:00:16.1: [8086:a361] type 00 class 0x078000 Aug 13 07:20:46.022426 kernel: pci 0000:00:16.1: reg 0x10: [mem 0x95519000-0x95519fff 64bit] Aug 13 07:20:46.022484 kernel: pci 0000:00:16.1: PME# supported from D3hot Aug 13 07:20:46.022538 kernel: pci 0000:00:16.4: [8086:a364] type 00 class 0x078000 Aug 13 07:20:46.022587 kernel: pci 0000:00:16.4: reg 0x10: [mem 0x95518000-0x95518fff 64bit] Aug 13 07:20:46.022636 kernel: pci 0000:00:16.4: PME# supported from D3hot Aug 13 07:20:46.022693 kernel: pci 0000:00:17.0: [8086:a352] type 00 class 0x010601 Aug 13 07:20:46.022743 kernel: pci 0000:00:17.0: reg 0x10: [mem 0x95510000-0x95511fff] Aug 13 07:20:46.022791 kernel: pci 0000:00:17.0: reg 0x14: [mem 0x95517000-0x955170ff] Aug 13 07:20:46.022840 kernel: pci 0000:00:17.0: reg 0x18: [io 0x6050-0x6057] Aug 13 07:20:46.022888 kernel: pci 0000:00:17.0: reg 0x1c: [io 0x6040-0x6043] Aug 13 07:20:46.022938 kernel: pci 0000:00:17.0: reg 0x20: [io 0x6020-0x603f] Aug 13 07:20:46.022986 kernel: pci 0000:00:17.0: reg 0x24: [mem 0x95516000-0x955167ff] Aug 13 07:20:46.023038 kernel: pci 0000:00:17.0: PME# supported from D3hot Aug 13 07:20:46.023134 kernel: pci 0000:00:1b.0: [8086:a340] type 01 class 0x060400 Aug 13 07:20:46.023187 kernel: pci 0000:00:1b.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023242 kernel: pci 0000:00:1b.4: [8086:a32c] type 01 class 0x060400 Aug 13 07:20:46.023291 kernel: pci 0000:00:1b.4: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023347 kernel: pci 0000:00:1b.5: [8086:a32d] type 01 class 0x060400 Aug 13 07:20:46.023397 kernel: pci 0000:00:1b.5: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023450 kernel: pci 0000:00:1c.0: [8086:a338] type 01 class 0x060400 Aug 13 07:20:46.023500 kernel: pci 0000:00:1c.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023555 kernel: pci 0000:00:1c.3: [8086:a33b] type 01 class 0x060400 Aug 13 07:20:46.023607 kernel: pci 0000:00:1c.3: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.023663 kernel: pci 0000:00:1e.0: [8086:a328] type 00 class 0x078000 Aug 13 07:20:46.023712 kernel: pci 0000:00:1e.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Aug 13 07:20:46.023766 kernel: pci 0000:00:1f.0: [8086:a309] type 00 class 0x060100 Aug 13 07:20:46.023818 kernel: pci 0000:00:1f.4: [8086:a323] type 00 class 0x0c0500 Aug 13 07:20:46.023869 kernel: pci 0000:00:1f.4: reg 0x10: [mem 0x95514000-0x955140ff 64bit] Aug 13 07:20:46.023920 kernel: pci 0000:00:1f.4: reg 0x20: [io 0xefa0-0xefbf] Aug 13 07:20:46.023974 kernel: pci 0000:00:1f.5: [8086:a324] type 00 class 0x0c8000 Aug 13 07:20:46.024023 kernel: pci 0000:00:1f.5: reg 0x10: [mem 0xfe010000-0xfe010fff] Aug 13 07:20:46.024105 kernel: pci 0000:01:00.0: [15b3:1015] type 00 class 0x020000 Aug 13 07:20:46.024174 kernel: pci 0000:01:00.0: reg 0x10: [mem 0x92000000-0x93ffffff 64bit pref] Aug 13 07:20:46.024225 kernel: pci 0000:01:00.0: reg 0x30: [mem 0x95200000-0x952fffff pref] Aug 13 07:20:46.024275 kernel: pci 0000:01:00.0: PME# supported from D3cold Aug 13 07:20:46.024329 kernel: pci 0000:01:00.0: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Aug 13 07:20:46.024380 kernel: pci 0000:01:00.0: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 07:20:46.024437 kernel: pci 0000:01:00.1: [15b3:1015] type 00 class 0x020000 Aug 13 07:20:46.024488 kernel: pci 0000:01:00.1: reg 0x10: [mem 0x90000000-0x91ffffff 64bit pref] Aug 13 07:20:46.024540 kernel: pci 0000:01:00.1: reg 0x30: [mem 0x95100000-0x951fffff pref] Aug 13 07:20:46.024590 kernel: pci 0000:01:00.1: PME# supported from D3cold Aug 13 07:20:46.024641 kernel: pci 0000:01:00.1: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Aug 13 07:20:46.024694 kernel: pci 0000:01:00.1: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 07:20:46.024745 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Aug 13 07:20:46.024794 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Aug 13 07:20:46.024844 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Aug 13 07:20:46.024892 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Aug 13 07:20:46.024950 kernel: pci 0000:03:00.0: working around ROM BAR overlap defect Aug 13 07:20:46.025002 kernel: pci 0000:03:00.0: [8086:1533] type 00 class 0x020000 Aug 13 07:20:46.025055 kernel: pci 0000:03:00.0: reg 0x10: [mem 0x95400000-0x9547ffff] Aug 13 07:20:46.025141 kernel: pci 0000:03:00.0: reg 0x18: [io 0x5000-0x501f] Aug 13 07:20:46.025191 kernel: pci 0000:03:00.0: reg 0x1c: [mem 0x95480000-0x95483fff] Aug 13 07:20:46.025241 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.025291 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Aug 13 07:20:46.025341 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Aug 13 07:20:46.025389 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Aug 13 07:20:46.025445 kernel: pci 0000:04:00.0: working around ROM BAR overlap defect Aug 13 07:20:46.025500 kernel: pci 0000:04:00.0: [8086:1533] type 00 class 0x020000 Aug 13 07:20:46.025550 kernel: pci 0000:04:00.0: reg 0x10: [mem 0x95300000-0x9537ffff] Aug 13 07:20:46.025600 kernel: pci 0000:04:00.0: reg 0x18: [io 0x4000-0x401f] Aug 13 07:20:46.025649 kernel: pci 0000:04:00.0: reg 0x1c: [mem 0x95380000-0x95383fff] Aug 13 07:20:46.025700 kernel: pci 0000:04:00.0: PME# supported from D0 D3hot D3cold Aug 13 07:20:46.025750 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Aug 13 07:20:46.025800 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Aug 13 07:20:46.025851 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Aug 13 07:20:46.025901 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Aug 13 07:20:46.025957 kernel: pci 0000:06:00.0: [1a03:1150] type 01 class 0x060400 Aug 13 07:20:46.026008 kernel: pci 0000:06:00.0: enabling Extended Tags Aug 13 07:20:46.026058 kernel: pci 0000:06:00.0: supports D1 D2 Aug 13 07:20:46.026151 kernel: pci 0000:06:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 07:20:46.026202 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Aug 13 07:20:46.026253 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.026302 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.026354 kernel: pci_bus 0000:07: extended config space not accessible Aug 13 07:20:46.026412 kernel: pci 0000:07:00.0: [1a03:2000] type 00 class 0x030000 Aug 13 07:20:46.026464 kernel: pci 0000:07:00.0: reg 0x10: [mem 0x94000000-0x94ffffff] Aug 13 07:20:46.026517 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x95000000-0x9501ffff] Aug 13 07:20:46.026571 kernel: pci 0000:07:00.0: reg 0x18: [io 0x3000-0x307f] Aug 13 07:20:46.026622 kernel: pci 0000:07:00.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Aug 13 07:20:46.026677 kernel: pci 0000:07:00.0: supports D1 D2 Aug 13 07:20:46.026729 kernel: pci 0000:07:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 07:20:46.026781 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Aug 13 07:20:46.026832 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.026882 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.026891 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 0 Aug 13 07:20:46.026897 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 1 Aug 13 07:20:46.026904 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 0 Aug 13 07:20:46.026910 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 0 Aug 13 07:20:46.026916 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 0 Aug 13 07:20:46.026921 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 0 Aug 13 07:20:46.026927 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 0 Aug 13 07:20:46.026932 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 0 Aug 13 07:20:46.026938 kernel: iommu: Default domain type: Translated Aug 13 07:20:46.026944 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Aug 13 07:20:46.026949 kernel: PCI: Using ACPI for IRQ routing Aug 13 07:20:46.026956 kernel: PCI: pci_cache_line_size set to 64 bytes Aug 13 07:20:46.026961 kernel: e820: reserve RAM buffer [mem 0x00099800-0x0009ffff] Aug 13 07:20:46.026967 kernel: e820: reserve RAM buffer [mem 0x819c7000-0x83ffffff] Aug 13 07:20:46.026973 kernel: e820: reserve RAM buffer [mem 0x8afce000-0x8bffffff] Aug 13 07:20:46.026979 kernel: e820: reserve RAM buffer [mem 0x8c23c000-0x8fffffff] Aug 13 07:20:46.026984 kernel: e820: reserve RAM buffer [mem 0x8ef00000-0x8fffffff] Aug 13 07:20:46.026990 kernel: e820: reserve RAM buffer [mem 0x86f000000-0x86fffffff] Aug 13 07:20:46.027042 kernel: pci 0000:07:00.0: vgaarb: setting as boot VGA device Aug 13 07:20:46.027122 kernel: pci 0000:07:00.0: vgaarb: bridge control possible Aug 13 07:20:46.027197 kernel: pci 0000:07:00.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Aug 13 07:20:46.027206 kernel: vgaarb: loaded Aug 13 07:20:46.027212 kernel: clocksource: Switched to clocksource tsc-early Aug 13 07:20:46.027217 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 07:20:46.027223 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 07:20:46.027229 kernel: pnp: PnP ACPI init Aug 13 07:20:46.027281 kernel: system 00:00: [mem 0x40000000-0x403fffff] has been reserved Aug 13 07:20:46.027332 kernel: pnp 00:02: [dma 0 disabled] Aug 13 07:20:46.027383 kernel: pnp 00:03: [dma 0 disabled] Aug 13 07:20:46.027432 kernel: system 00:04: [io 0x0680-0x069f] has been reserved Aug 13 07:20:46.027477 kernel: system 00:04: [io 0x164e-0x164f] has been reserved Aug 13 07:20:46.027526 kernel: system 00:05: [mem 0xfed10000-0xfed17fff] has been reserved Aug 13 07:20:46.027571 kernel: system 00:05: [mem 0xfed18000-0xfed18fff] has been reserved Aug 13 07:20:46.027615 kernel: system 00:05: [mem 0xfed19000-0xfed19fff] has been reserved Aug 13 07:20:46.027663 kernel: system 00:05: [mem 0xe0000000-0xefffffff] has been reserved Aug 13 07:20:46.027707 kernel: system 00:05: [mem 0xfed20000-0xfed3ffff] has been reserved Aug 13 07:20:46.027752 kernel: system 00:05: [mem 0xfed90000-0xfed93fff] could not be reserved Aug 13 07:20:46.027796 kernel: system 00:05: [mem 0xfed45000-0xfed8ffff] has been reserved Aug 13 07:20:46.027842 kernel: system 00:05: [mem 0xfee00000-0xfeefffff] could not be reserved Aug 13 07:20:46.027891 kernel: system 00:06: [io 0x1800-0x18fe] could not be reserved Aug 13 07:20:46.027940 kernel: system 00:06: [mem 0xfd000000-0xfd69ffff] has been reserved Aug 13 07:20:46.027987 kernel: system 00:06: [mem 0xfd6c0000-0xfd6cffff] has been reserved Aug 13 07:20:46.028032 kernel: system 00:06: [mem 0xfd6f0000-0xfdffffff] has been reserved Aug 13 07:20:46.028099 kernel: system 00:06: [mem 0xfe000000-0xfe01ffff] could not be reserved Aug 13 07:20:46.028163 kernel: system 00:06: [mem 0xfe200000-0xfe7fffff] has been reserved Aug 13 07:20:46.028208 kernel: system 00:06: [mem 0xff000000-0xffffffff] has been reserved Aug 13 07:20:46.028256 kernel: system 00:07: [io 0x2000-0x20fe] has been reserved Aug 13 07:20:46.028265 kernel: pnp: PnP ACPI: found 9 devices Aug 13 07:20:46.028273 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Aug 13 07:20:46.028279 kernel: NET: Registered PF_INET protocol family Aug 13 07:20:46.028284 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 07:20:46.028290 kernel: tcp_listen_portaddr_hash hash table entries: 16384 (order: 6, 262144 bytes, linear) Aug 13 07:20:46.028296 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 07:20:46.028302 kernel: TCP established hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 07:20:46.028307 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 07:20:46.028313 kernel: TCP: Hash tables configured (established 262144 bind 65536) Aug 13 07:20:46.028318 kernel: UDP hash table entries: 16384 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.028325 kernel: UDP-Lite hash table entries: 16384 (order: 7, 524288 bytes, linear) Aug 13 07:20:46.028331 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 07:20:46.028337 kernel: NET: Registered PF_XDP protocol family Aug 13 07:20:46.028387 kernel: pci 0000:00:15.0: BAR 0: assigned [mem 0x95515000-0x95515fff 64bit] Aug 13 07:20:46.028437 kernel: pci 0000:00:15.1: BAR 0: assigned [mem 0x9551b000-0x9551bfff 64bit] Aug 13 07:20:46.028487 kernel: pci 0000:00:1e.0: BAR 0: assigned [mem 0x9551c000-0x9551cfff 64bit] Aug 13 07:20:46.028539 kernel: pci 0000:01:00.0: BAR 7: no space for [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028590 kernel: pci 0000:01:00.0: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028643 kernel: pci 0000:01:00.1: BAR 7: no space for [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028695 kernel: pci 0000:01:00.1: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Aug 13 07:20:46.028744 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Aug 13 07:20:46.028794 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Aug 13 07:20:46.028843 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Aug 13 07:20:46.028893 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Aug 13 07:20:46.028946 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Aug 13 07:20:46.028995 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Aug 13 07:20:46.029047 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Aug 13 07:20:46.029134 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Aug 13 07:20:46.029185 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Aug 13 07:20:46.029233 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Aug 13 07:20:46.029283 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Aug 13 07:20:46.029336 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Aug 13 07:20:46.029387 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.029437 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.029487 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Aug 13 07:20:46.029537 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Aug 13 07:20:46.029586 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Aug 13 07:20:46.029632 kernel: pci_bus 0000:00: Some PCI device resources are unassigned, try booting with pci=realloc Aug 13 07:20:46.029675 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Aug 13 07:20:46.029719 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Aug 13 07:20:46.029764 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Aug 13 07:20:46.029807 kernel: pci_bus 0000:00: resource 7 [mem 0x90000000-0xdfffffff window] Aug 13 07:20:46.029849 kernel: pci_bus 0000:00: resource 8 [mem 0xfc800000-0xfe7fffff window] Aug 13 07:20:46.029902 kernel: pci_bus 0000:01: resource 1 [mem 0x95100000-0x952fffff] Aug 13 07:20:46.029947 kernel: pci_bus 0000:01: resource 2 [mem 0x90000000-0x93ffffff 64bit pref] Aug 13 07:20:46.029999 kernel: pci_bus 0000:03: resource 0 [io 0x5000-0x5fff] Aug 13 07:20:46.030047 kernel: pci_bus 0000:03: resource 1 [mem 0x95400000-0x954fffff] Aug 13 07:20:46.030143 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Aug 13 07:20:46.030188 kernel: pci_bus 0000:04: resource 1 [mem 0x95300000-0x953fffff] Aug 13 07:20:46.030236 kernel: pci_bus 0000:06: resource 0 [io 0x3000-0x3fff] Aug 13 07:20:46.030281 kernel: pci_bus 0000:06: resource 1 [mem 0x94000000-0x950fffff] Aug 13 07:20:46.030329 kernel: pci_bus 0000:07: resource 0 [io 0x3000-0x3fff] Aug 13 07:20:46.030375 kernel: pci_bus 0000:07: resource 1 [mem 0x94000000-0x950fffff] Aug 13 07:20:46.030385 kernel: PCI: CLS 64 bytes, default 64 Aug 13 07:20:46.030391 kernel: DMAR: No ATSR found Aug 13 07:20:46.030396 kernel: DMAR: No SATC found Aug 13 07:20:46.030402 kernel: DMAR: dmar0: Using Queued invalidation Aug 13 07:20:46.030452 kernel: pci 0000:00:00.0: Adding to iommu group 0 Aug 13 07:20:46.030501 kernel: pci 0000:00:01.0: Adding to iommu group 1 Aug 13 07:20:46.030552 kernel: pci 0000:00:08.0: Adding to iommu group 2 Aug 13 07:20:46.030602 kernel: pci 0000:00:12.0: Adding to iommu group 3 Aug 13 07:20:46.030651 kernel: pci 0000:00:14.0: Adding to iommu group 4 Aug 13 07:20:46.030704 kernel: pci 0000:00:14.2: Adding to iommu group 4 Aug 13 07:20:46.030752 kernel: pci 0000:00:15.0: Adding to iommu group 5 Aug 13 07:20:46.030801 kernel: pci 0000:00:15.1: Adding to iommu group 5 Aug 13 07:20:46.030849 kernel: pci 0000:00:16.0: Adding to iommu group 6 Aug 13 07:20:46.030899 kernel: pci 0000:00:16.1: Adding to iommu group 6 Aug 13 07:20:46.030947 kernel: pci 0000:00:16.4: Adding to iommu group 6 Aug 13 07:20:46.030997 kernel: pci 0000:00:17.0: Adding to iommu group 7 Aug 13 07:20:46.031045 kernel: pci 0000:00:1b.0: Adding to iommu group 8 Aug 13 07:20:46.031144 kernel: pci 0000:00:1b.4: Adding to iommu group 9 Aug 13 07:20:46.031193 kernel: pci 0000:00:1b.5: Adding to iommu group 10 Aug 13 07:20:46.031243 kernel: pci 0000:00:1c.0: Adding to iommu group 11 Aug 13 07:20:46.031291 kernel: pci 0000:00:1c.3: Adding to iommu group 12 Aug 13 07:20:46.031340 kernel: pci 0000:00:1e.0: Adding to iommu group 13 Aug 13 07:20:46.031390 kernel: pci 0000:00:1f.0: Adding to iommu group 14 Aug 13 07:20:46.031439 kernel: pci 0000:00:1f.4: Adding to iommu group 14 Aug 13 07:20:46.031488 kernel: pci 0000:00:1f.5: Adding to iommu group 14 Aug 13 07:20:46.031542 kernel: pci 0000:01:00.0: Adding to iommu group 1 Aug 13 07:20:46.031593 kernel: pci 0000:01:00.1: Adding to iommu group 1 Aug 13 07:20:46.031644 kernel: pci 0000:03:00.0: Adding to iommu group 15 Aug 13 07:20:46.031696 kernel: pci 0000:04:00.0: Adding to iommu group 16 Aug 13 07:20:46.031746 kernel: pci 0000:06:00.0: Adding to iommu group 17 Aug 13 07:20:46.031800 kernel: pci 0000:07:00.0: Adding to iommu group 17 Aug 13 07:20:46.031808 kernel: DMAR: Intel(R) Virtualization Technology for Directed I/O Aug 13 07:20:46.031814 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Aug 13 07:20:46.031822 kernel: software IO TLB: mapped [mem 0x0000000086fce000-0x000000008afce000] (64MB) Aug 13 07:20:46.031828 kernel: RAPL PMU: API unit is 2^-32 Joules, 3 fixed counters, 655360 ms ovfl timer Aug 13 07:20:46.031833 kernel: RAPL PMU: hw unit of domain pp0-core 2^-14 Joules Aug 13 07:20:46.031839 kernel: RAPL PMU: hw unit of domain package 2^-14 Joules Aug 13 07:20:46.031845 kernel: RAPL PMU: hw unit of domain dram 2^-14 Joules Aug 13 07:20:46.031899 kernel: platform rtc_cmos: registered platform RTC device (no PNP device found) Aug 13 07:20:46.031908 kernel: Initialise system trusted keyrings Aug 13 07:20:46.031914 kernel: workingset: timestamp_bits=39 max_order=23 bucket_order=0 Aug 13 07:20:46.031921 kernel: Key type asymmetric registered Aug 13 07:20:46.031927 kernel: Asymmetric key parser 'x509' registered Aug 13 07:20:46.031932 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Aug 13 07:20:46.031938 kernel: io scheduler mq-deadline registered Aug 13 07:20:46.031944 kernel: io scheduler kyber registered Aug 13 07:20:46.031949 kernel: io scheduler bfq registered Aug 13 07:20:46.031998 kernel: pcieport 0000:00:01.0: PME: Signaling with IRQ 121 Aug 13 07:20:46.032048 kernel: pcieport 0000:00:1b.0: PME: Signaling with IRQ 122 Aug 13 07:20:46.032146 kernel: pcieport 0000:00:1b.4: PME: Signaling with IRQ 123 Aug 13 07:20:46.032199 kernel: pcieport 0000:00:1b.5: PME: Signaling with IRQ 124 Aug 13 07:20:46.032248 kernel: pcieport 0000:00:1c.0: PME: Signaling with IRQ 125 Aug 13 07:20:46.032298 kernel: pcieport 0000:00:1c.3: PME: Signaling with IRQ 126 Aug 13 07:20:46.032351 kernel: thermal LNXTHERM:00: registered as thermal_zone0 Aug 13 07:20:46.032359 kernel: ACPI: thermal: Thermal Zone [TZ00] (28 C) Aug 13 07:20:46.032365 kernel: ERST: Error Record Serialization Table (ERST) support is initialized. Aug 13 07:20:46.032371 kernel: pstore: Using crash dump compression: deflate Aug 13 07:20:46.032377 kernel: pstore: Registered erst as persistent store backend Aug 13 07:20:46.032384 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Aug 13 07:20:46.032390 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 07:20:46.032395 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Aug 13 07:20:46.032401 kernel: 00:03: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Aug 13 07:20:46.032407 kernel: hpet_acpi_add: no address or irqs in _CRS Aug 13 07:20:46.032456 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x1B, rev-id 16) Aug 13 07:20:46.032465 kernel: i8042: PNP: No PS/2 controller found. Aug 13 07:20:46.032509 kernel: rtc_cmos rtc_cmos: RTC can wake from S4 Aug 13 07:20:46.032557 kernel: rtc_cmos rtc_cmos: registered as rtc0 Aug 13 07:20:46.032601 kernel: rtc_cmos rtc_cmos: setting system clock to 2025-08-13T07:20:44 UTC (1755069644) Aug 13 07:20:46.032647 kernel: rtc_cmos rtc_cmos: alarms up to one month, y3k, 114 bytes nvram Aug 13 07:20:46.032655 kernel: intel_pstate: Intel P-state driver initializing Aug 13 07:20:46.032661 kernel: intel_pstate: Disabling energy efficiency optimization Aug 13 07:20:46.032666 kernel: intel_pstate: HWP enabled Aug 13 07:20:46.032672 kernel: vesafb: mode is 1024x768x8, linelength=1024, pages=0 Aug 13 07:20:46.032678 kernel: vesafb: scrolling: redraw Aug 13 07:20:46.032685 kernel: vesafb: Pseudocolor: size=0:8:8:8, shift=0:0:0:0 Aug 13 07:20:46.032690 kernel: vesafb: framebuffer at 0x94000000, mapped to 0x000000005212a1e2, using 768k, total 768k Aug 13 07:20:46.032696 kernel: Console: switching to colour frame buffer device 128x48 Aug 13 07:20:46.032702 kernel: fb0: VESA VGA frame buffer device Aug 13 07:20:46.032707 kernel: NET: Registered PF_INET6 protocol family Aug 13 07:20:46.032713 kernel: Segment Routing with IPv6 Aug 13 07:20:46.032719 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 07:20:46.032724 kernel: NET: Registered PF_PACKET protocol family Aug 13 07:20:46.032730 kernel: Key type dns_resolver registered Aug 13 07:20:46.032736 kernel: microcode: Current revision: 0x00000102 Aug 13 07:20:46.032742 kernel: microcode: Microcode Update Driver: v2.2. Aug 13 07:20:46.032747 kernel: IPI shorthand broadcast: enabled Aug 13 07:20:46.032753 kernel: sched_clock: Marking stable (1561000725, 1379333307)->(4403206493, -1462872461) Aug 13 07:20:46.032759 kernel: registered taskstats version 1 Aug 13 07:20:46.032764 kernel: Loading compiled-in X.509 certificates Aug 13 07:20:46.032770 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.100-flatcar: 264e720147fa8df9744bb9dc1c08171c0cb20041' Aug 13 07:20:46.032775 kernel: Key type .fscrypt registered Aug 13 07:20:46.032781 kernel: Key type fscrypt-provisioning registered Aug 13 07:20:46.032788 kernel: ima: Allocated hash algorithm: sha1 Aug 13 07:20:46.032793 kernel: ima: No architecture policies found Aug 13 07:20:46.032799 kernel: clk: Disabling unused clocks Aug 13 07:20:46.032805 kernel: Freeing unused kernel image (initmem) memory: 42876K Aug 13 07:20:46.032810 kernel: Write protecting the kernel read-only data: 36864k Aug 13 07:20:46.032816 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Aug 13 07:20:46.032822 kernel: Run /init as init process Aug 13 07:20:46.032827 kernel: with arguments: Aug 13 07:20:46.032833 kernel: /init Aug 13 07:20:46.032840 kernel: with environment: Aug 13 07:20:46.032845 kernel: HOME=/ Aug 13 07:20:46.032851 kernel: TERM=linux Aug 13 07:20:46.032856 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 07:20:46.032863 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 07:20:46.032870 systemd[1]: Detected architecture x86-64. Aug 13 07:20:46.032876 systemd[1]: Running in initrd. Aug 13 07:20:46.032883 systemd[1]: No hostname configured, using default hostname. Aug 13 07:20:46.032889 systemd[1]: Hostname set to . Aug 13 07:20:46.032895 systemd[1]: Initializing machine ID from random generator. Aug 13 07:20:46.032901 systemd[1]: Queued start job for default target initrd.target. Aug 13 07:20:46.032907 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:20:46.032913 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:20:46.032919 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 07:20:46.032925 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:20:46.032932 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 07:20:46.032938 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 07:20:46.032944 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 07:20:46.032951 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 07:20:46.032957 kernel: tsc: Refined TSC clocksource calibration: 3407.999 MHz Aug 13 07:20:46.032963 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd336761, max_idle_ns: 440795243819 ns Aug 13 07:20:46.032968 kernel: clocksource: Switched to clocksource tsc Aug 13 07:20:46.032975 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:20:46.032981 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:20:46.032987 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:20:46.032993 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:20:46.032999 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:20:46.033005 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:20:46.033011 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:20:46.033016 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:20:46.033022 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 07:20:46.033029 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Aug 13 07:20:46.033035 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:20:46.033041 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:20:46.033047 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:20:46.033053 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:20:46.033059 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 07:20:46.033068 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:20:46.033074 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 07:20:46.033101 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 07:20:46.033107 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:20:46.033137 systemd-journald[267]: Collecting audit messages is disabled. Aug 13 07:20:46.033151 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:20:46.033158 systemd-journald[267]: Journal started Aug 13 07:20:46.033172 systemd-journald[267]: Runtime Journal (/run/log/journal/59478b815246437cb8b89f1def936f76) is 8.0M, max 639.9M, 631.9M free. Aug 13 07:20:46.047556 systemd-modules-load[268]: Inserted module 'overlay' Aug 13 07:20:46.068067 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:20:46.088082 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:20:46.100381 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 07:20:46.100550 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:20:46.100690 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 07:20:46.101655 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 07:20:46.101989 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:20:46.143099 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 07:20:46.161650 systemd-modules-load[268]: Inserted module 'br_netfilter' Aug 13 07:20:46.242422 kernel: Bridge firewalling registered Aug 13 07:20:46.161992 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:20:46.259488 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:20:46.280369 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 07:20:46.301503 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:20:46.347291 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:20:46.347733 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:20:46.379348 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:20:46.385187 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:20:46.386445 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:20:46.408046 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:20:46.436369 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 07:20:46.457181 dracut-cmdline[304]: dracut-dracut-053 Aug 13 07:20:46.457181 dracut-cmdline[304]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:20:46.522127 kernel: SCSI subsystem initialized Aug 13 07:20:46.447866 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:20:46.552437 kernel: Loading iSCSI transport class v2.0-870. Aug 13 07:20:46.552450 kernel: iscsi: registered transport (tcp) Aug 13 07:20:46.471364 systemd-resolved[323]: Positive Trust Anchors: Aug 13 07:20:46.582935 kernel: iscsi: registered transport (qla4xxx) Aug 13 07:20:46.471372 systemd-resolved[323]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:20:46.621326 kernel: QLogic iSCSI HBA Driver Aug 13 07:20:46.471403 systemd-resolved[323]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:20:46.473504 systemd-resolved[323]: Defaulting to hostname 'linux'. Aug 13 07:20:46.474205 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:20:46.540172 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:20:46.618099 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 07:20:46.642387 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 07:20:46.821230 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 07:20:46.821255 kernel: device-mapper: uevent: version 1.0.3 Aug 13 07:20:46.840981 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Aug 13 07:20:46.900139 kernel: raid6: avx2x4 gen() 53374 MB/s Aug 13 07:20:46.932141 kernel: raid6: avx2x2 gen() 53939 MB/s Aug 13 07:20:46.968540 kernel: raid6: avx2x1 gen() 45218 MB/s Aug 13 07:20:46.968559 kernel: raid6: using algorithm avx2x2 gen() 53939 MB/s Aug 13 07:20:47.015599 kernel: raid6: .... xor() 31363 MB/s, rmw enabled Aug 13 07:20:47.015617 kernel: raid6: using avx2x2 recovery algorithm Aug 13 07:20:47.057127 kernel: xor: automatically using best checksumming function avx Aug 13 07:20:47.171076 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 07:20:47.176428 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:20:47.206405 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:20:47.213287 systemd-udevd[494]: Using default interface naming scheme 'v255'. Aug 13 07:20:47.217195 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:20:47.258461 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 07:20:47.320245 dracut-pre-trigger[507]: rd.md=0: removing MD RAID activation Aug 13 07:20:47.340880 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:20:47.356418 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:20:47.421579 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:20:47.454334 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 07:20:47.454416 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 07:20:47.480072 kernel: cryptd: max_cpu_qlen set to 1000 Aug 13 07:20:47.483252 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 07:20:47.509174 kernel: libata version 3.00 loaded. Aug 13 07:20:47.509188 kernel: PTP clock support registered Aug 13 07:20:47.509196 kernel: ACPI: bus type USB registered Aug 13 07:20:47.484615 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:20:47.576177 kernel: usbcore: registered new interface driver usbfs Aug 13 07:20:47.576194 kernel: usbcore: registered new interface driver hub Aug 13 07:20:47.576202 kernel: usbcore: registered new device driver usb Aug 13 07:20:47.576209 kernel: AVX2 version of gcm_enc/dec engaged. Aug 13 07:20:47.484647 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:20:47.598176 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:20:47.739398 kernel: AES CTR mode by8 optimization enabled Aug 13 07:20:47.739415 kernel: ahci 0000:00:17.0: version 3.0 Aug 13 07:20:47.739510 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Aug 13 07:20:47.739519 kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 7 ports 6 Gbps 0x7f impl SATA mode Aug 13 07:20:47.739588 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Aug 13 07:20:47.739596 kernel: ahci 0000:00:17.0: flags: 64bit ncq sntf clo only pio slum part ems deso sadm sds apst Aug 13 07:20:47.739659 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Aug 13 07:20:47.739722 kernel: igb 0000:03:00.0: added PHC on eth0 Aug 13 07:20:47.739790 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 1 Aug 13 07:20:47.739851 kernel: igb 0000:03:00.0: Intel(R) Gigabit Ethernet Network Connection Aug 13 07:20:47.739914 kernel: xhci_hcd 0000:00:14.0: hcc params 0x200077c1 hci version 0x110 quirks 0x0000000000009810 Aug 13 07:20:47.739974 kernel: igb 0000:03:00.0: eth0: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:ef:be Aug 13 07:20:47.740038 kernel: scsi host0: ahci Aug 13 07:20:47.740107 kernel: scsi host1: ahci Aug 13 07:20:47.740168 kernel: scsi host2: ahci Aug 13 07:20:47.740226 kernel: scsi host3: ahci Aug 13 07:20:47.740287 kernel: scsi host4: ahci Aug 13 07:20:47.740343 kernel: scsi host5: ahci Aug 13 07:20:47.740402 kernel: scsi host6: ahci Aug 13 07:20:47.740461 kernel: ata1: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516100 irq 127 Aug 13 07:20:47.740469 kernel: ata2: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516180 irq 127 Aug 13 07:20:47.740477 kernel: ata3: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516200 irq 127 Aug 13 07:20:47.740484 kernel: ata4: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516280 irq 127 Aug 13 07:20:47.740491 kernel: ata5: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516300 irq 127 Aug 13 07:20:47.740498 kernel: ata6: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516380 irq 127 Aug 13 07:20:47.740505 kernel: ata7: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516400 irq 127 Aug 13 07:20:47.740514 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Aug 13 07:20:47.740576 kernel: igb 0000:03:00.0: eth0: PBA No: 010000-000 Aug 13 07:20:47.740639 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 2 Aug 13 07:20:47.740699 kernel: igb 0000:03:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Aug 13 07:20:47.740761 kernel: xhci_hcd 0000:00:14.0: Host supports USB 3.1 Enhanced SuperSpeed Aug 13 07:20:47.614153 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:20:48.282095 kernel: igb 0000:04:00.0: added PHC on eth1 Aug 13 07:20:48.282182 kernel: hub 1-0:1.0: USB hub found Aug 13 07:20:48.282258 kernel: igb 0000:04:00.0: Intel(R) Gigabit Ethernet Network Connection Aug 13 07:20:48.282325 kernel: hub 1-0:1.0: 16 ports detected Aug 13 07:20:48.282387 kernel: igb 0000:04:00.0: eth1: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:ef:bf Aug 13 07:20:48.282452 kernel: hub 2-0:1.0: USB hub found Aug 13 07:20:48.282517 kernel: igb 0000:04:00.0: eth1: PBA No: 010000-000 Aug 13 07:20:48.282582 kernel: hub 2-0:1.0: 10 ports detected Aug 13 07:20:48.282643 kernel: igb 0000:04:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Aug 13 07:20:48.282706 kernel: ata1: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Aug 13 07:20:48.282715 kernel: ata5: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282722 kernel: ata1.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Aug 13 07:20:48.282729 kernel: ata4: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282736 kernel: ata6: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282743 kernel: usb 1-14: new high-speed USB device number 2 using xhci_hcd Aug 13 07:20:48.282760 kernel: ata3: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282767 kernel: ata7: SATA link down (SStatus 0 SControl 300) Aug 13 07:20:48.282774 kernel: ata1.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Aug 13 07:20:48.282781 kernel: ata2: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Aug 13 07:20:47.614195 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:20:48.357167 kernel: ata2.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Aug 13 07:20:48.357179 kernel: mlx5_core 0000:01:00.0: firmware version: 14.31.1014 Aug 13 07:20:48.357266 kernel: ata2.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Aug 13 07:20:48.357278 kernel: mlx5_core 0000:01:00.0: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Aug 13 07:20:48.357346 kernel: ata1.00: Features: NCQ-prio Aug 13 07:20:48.205165 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:20:48.437302 kernel: hub 1-14:1.0: USB hub found Aug 13 07:20:48.437391 kernel: ata2.00: Features: NCQ-prio Aug 13 07:20:48.437403 kernel: hub 1-14:1.0: 4 ports detected Aug 13 07:20:48.437471 kernel: ata1.00: configured for UDMA/133 Aug 13 07:20:48.437480 kernel: scsi 0:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Aug 13 07:20:48.387227 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:20:48.465652 kernel: ata2.00: configured for UDMA/133 Aug 13 07:20:48.465663 kernel: scsi 1:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Aug 13 07:20:48.437473 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 07:20:48.466151 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:20:48.483470 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:20:48.483513 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:20:48.500104 kernel: igb 0000:03:00.0 eno1: renamed from eth0 Aug 13 07:20:48.501144 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 07:20:48.984193 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:48.984211 kernel: igb 0000:04:00.0 eno2: renamed from eth1 Aug 13 07:20:48.984307 kernel: ata2.00: Enabling discard_zeroes_data Aug 13 07:20:48.984320 kernel: sd 0:0:0:0: [sdb] 937703088 512-byte logical blocks: (480 GB/447 GiB) Aug 13 07:20:48.984395 kernel: sd 1:0:0:0: [sda] 937703088 512-byte logical blocks: (480 GB/447 GiB) Aug 13 07:20:48.984463 kernel: sd 0:0:0:0: [sdb] 4096-byte physical blocks Aug 13 07:20:48.984525 kernel: sd 0:0:0:0: [sdb] Write Protect is off Aug 13 07:20:48.984587 kernel: sd 1:0:0:0: [sda] 4096-byte physical blocks Aug 13 07:20:48.984647 kernel: sd 0:0:0:0: [sdb] Mode Sense: 00 3a 00 00 Aug 13 07:20:48.984709 kernel: sd 0:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Aug 13 07:20:48.984772 kernel: sd 0:0:0:0: [sdb] Preferred minimum I/O size 4096 bytes Aug 13 07:20:48.984834 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:48.984843 kernel: mlx5_core 0000:01:00.0: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Aug 13 07:20:48.984912 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 07:20:48.984921 kernel: sd 1:0:0:0: [sda] Write Protect is off Aug 13 07:20:48.984982 kernel: mlx5_core 0000:01:00.0: Port module event: module 0, Cable plugged Aug 13 07:20:48.985048 kernel: GPT:9289727 != 937703087 Aug 13 07:20:48.985056 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 07:20:48.985071 kernel: GPT:9289727 != 937703087 Aug 13 07:20:48.985079 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 07:20:48.985086 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:48.985093 kernel: sd 1:0:0:0: [sda] Mode Sense: 00 3a 00 00 Aug 13 07:20:48.985157 kernel: sd 0:0:0:0: [sdb] Attached SCSI disk Aug 13 07:20:48.985219 kernel: sd 1:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Aug 13 07:20:48.985279 kernel: BTRFS: device fsid 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad devid 1 transid 37 /dev/sdb3 scanned by (udev-worker) (558) Aug 13 07:20:48.985288 kernel: sd 1:0:0:0: [sda] Preferred minimum I/O size 4096 bytes Aug 13 07:20:48.985350 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sdb6 scanned by (udev-worker) (541) Aug 13 07:20:48.985358 kernel: usb 1-14.1: new low-speed USB device number 3 using xhci_hcd Aug 13 07:20:48.985464 kernel: ata2.00: Enabling discard_zeroes_data Aug 13 07:20:48.985473 kernel: sd 1:0:0:0: [sda] Attached SCSI disk Aug 13 07:20:48.985536 kernel: hid: raw HID events driver (C) Jiri Kosina Aug 13 07:20:48.958725 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:20:49.003407 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Micron_5300_MTFDDAK480TDT EFI-SYSTEM. Aug 13 07:20:49.074156 kernel: usbcore: registered new interface driver usbhid Aug 13 07:20:49.074169 kernel: usbhid: USB HID core driver Aug 13 07:20:49.074177 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.0/0003:0557:2419.0001/input/input0 Aug 13 07:20:49.054716 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Micron_5300_MTFDDAK480TDT ROOT. Aug 13 07:20:49.088235 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:20:49.281519 kernel: hid-generic 0003:0557:2419.0001: input,hidraw0: USB HID v1.00 Keyboard [HID 0557:2419] on usb-0000:00:14.0-14.1/input0 Aug 13 07:20:49.281653 kernel: mlx5_core 0000:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 07:20:49.281768 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.1/0003:0557:2419.0002/input/input1 Aug 13 07:20:49.281784 kernel: mlx5_core 0000:01:00.1: firmware version: 14.31.1014 Aug 13 07:20:49.281892 kernel: hid-generic 0003:0557:2419.0002: input,hidraw1: USB HID v1.00 Mouse [HID 0557:2419] on usb-0000:00:14.0-14.1/input1 Aug 13 07:20:49.282006 kernel: mlx5_core 0000:01:00.1: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Aug 13 07:20:49.113671 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Micron_5300_MTFDDAK480TDT USR-A. Aug 13 07:20:49.292158 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Micron_5300_MTFDDAK480TDT USR-A. Aug 13 07:20:49.320793 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Aug 13 07:20:49.363227 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 07:20:49.401188 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:49.401202 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:49.379556 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:20:49.425738 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:49.425794 disk-uuid[703]: Primary Header is updated. Aug 13 07:20:49.425794 disk-uuid[703]: Secondary Entries is updated. Aug 13 07:20:49.425794 disk-uuid[703]: Secondary Header is updated. Aug 13 07:20:49.468744 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:49.468755 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:49.489104 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:49.501460 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:20:49.600127 kernel: mlx5_core 0000:01:00.1: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Aug 13 07:20:49.624977 kernel: mlx5_core 0000:01:00.1: Port module event: module 1, Cable plugged Aug 13 07:20:49.927092 kernel: mlx5_core 0000:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 07:20:49.952129 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: renamed from eth0 Aug 13 07:20:49.978144 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: renamed from eth1 Aug 13 07:20:50.467646 kernel: ata1.00: Enabling discard_zeroes_data Aug 13 07:20:50.488131 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Aug 13 07:20:50.488179 disk-uuid[704]: The operation has completed successfully. Aug 13 07:20:50.524178 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 07:20:50.524244 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 07:20:50.558354 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 07:20:50.584236 sh[742]: Success Aug 13 07:20:50.594163 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Aug 13 07:20:50.633109 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 07:20:50.643023 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 07:20:50.661059 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 07:20:50.723157 kernel: BTRFS info (device dm-0): first mount of filesystem 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad Aug 13 07:20:50.723203 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:20:50.723236 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Aug 13 07:20:50.723258 kernel: BTRFS info (device dm-0): disabling log replay at mount time Aug 13 07:20:50.723278 kernel: BTRFS info (device dm-0): using free space tree Aug 13 07:20:50.723293 kernel: BTRFS info (device dm-0): enabling ssd optimizations Aug 13 07:20:50.796417 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 07:20:50.805486 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 07:20:50.812349 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 07:20:50.851573 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 07:20:50.893132 kernel: BTRFS info (device sdb6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:20:50.893170 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:20:50.906264 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:20:50.994887 kernel: BTRFS info (device sdb6): using free space tree Aug 13 07:20:50.994902 kernel: BTRFS info (device sdb6): enabling ssd optimizations Aug 13 07:20:50.994909 kernel: BTRFS info (device sdb6): auto enabling async discard Aug 13 07:20:50.994916 kernel: BTRFS info (device sdb6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:20:51.011425 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:20:51.011557 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 07:20:51.022560 systemd-networkd[922]: lo: Link UP Aug 13 07:20:51.022563 systemd-networkd[922]: lo: Gained carrier Aug 13 07:20:51.025133 systemd-networkd[922]: Enumeration completed Aug 13 07:20:51.026117 systemd-networkd[922]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.038390 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:20:51.054034 systemd-networkd[922]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.056594 systemd[1]: Reached target network.target - Network. Aug 13 07:20:51.084043 systemd-networkd[922]: enp1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.089527 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 07:20:51.156684 unknown[927]: fetched base config from "system" Aug 13 07:20:51.153933 ignition[927]: Ignition 2.19.0 Aug 13 07:20:51.156690 unknown[927]: fetched user config from "system" Aug 13 07:20:51.153937 ignition[927]: Stage: fetch-offline Aug 13 07:20:51.157798 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:20:51.153962 ignition[927]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:20:51.174218 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Aug 13 07:20:51.153968 ignition[927]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:20:51.183348 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 07:20:51.154027 ignition[927]: parsed url from cmdline: "" Aug 13 07:20:51.154029 ignition[927]: no config URL provided Aug 13 07:20:51.289249 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Aug 13 07:20:51.284438 systemd-networkd[922]: enp1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:20:51.154031 ignition[927]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 07:20:51.154055 ignition[927]: parsing config with SHA512: dbf51eb6cc0653fb3cc72ba56cd8c158fea4eda5e9d0bbcff0ccf5e44f0db0863474bfdbd18a4f57d5872bf9d2e499f2e0a83b1a3b2a27ebec96c34551325a17 Aug 13 07:20:51.156934 ignition[927]: fetch-offline: fetch-offline passed Aug 13 07:20:51.156937 ignition[927]: POST message to Packet Timeline Aug 13 07:20:51.156940 ignition[927]: POST Status error: resource requires networking Aug 13 07:20:51.156977 ignition[927]: Ignition finished successfully Aug 13 07:20:51.191331 ignition[942]: Ignition 2.19.0 Aug 13 07:20:51.191336 ignition[942]: Stage: kargs Aug 13 07:20:51.191463 ignition[942]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:20:51.191471 ignition[942]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:20:51.192172 ignition[942]: kargs: kargs passed Aug 13 07:20:51.192175 ignition[942]: POST message to Packet Timeline Aug 13 07:20:51.192187 ignition[942]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:20:51.192711 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:54753->[::1]:53: read: connection refused Aug 13 07:20:51.393534 ignition[942]: GET https://metadata.packet.net/metadata: attempt #2 Aug 13 07:20:51.394668 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:38785->[::1]:53: read: connection refused Aug 13 07:20:51.547201 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Aug 13 07:20:51.547940 systemd-networkd[922]: eno1: Link UP Aug 13 07:20:51.548116 systemd-networkd[922]: eno2: Link UP Aug 13 07:20:51.548288 systemd-networkd[922]: enp1s0f0np0: Link UP Aug 13 07:20:51.548468 systemd-networkd[922]: enp1s0f0np0: Gained carrier Aug 13 07:20:51.557332 systemd-networkd[922]: enp1s0f1np1: Link UP Aug 13 07:20:51.585235 systemd-networkd[922]: enp1s0f0np0: DHCPv4 address 145.40.90.243/31, gateway 145.40.90.242 acquired from 145.40.83.140 Aug 13 07:20:51.795119 ignition[942]: GET https://metadata.packet.net/metadata: attempt #3 Aug 13 07:20:51.796127 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:50675->[::1]:53: read: connection refused Aug 13 07:20:52.330728 systemd-networkd[922]: enp1s0f1np1: Gained carrier Aug 13 07:20:52.596664 ignition[942]: GET https://metadata.packet.net/metadata: attempt #4 Aug 13 07:20:52.597721 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:38445->[::1]:53: read: connection refused Aug 13 07:20:52.842692 systemd-networkd[922]: enp1s0f0np0: Gained IPv6LL Aug 13 07:20:53.930703 systemd-networkd[922]: enp1s0f1np1: Gained IPv6LL Aug 13 07:20:54.199342 ignition[942]: GET https://metadata.packet.net/metadata: attempt #5 Aug 13 07:20:54.200452 ignition[942]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:55213->[::1]:53: read: connection refused Aug 13 07:20:57.403964 ignition[942]: GET https://metadata.packet.net/metadata: attempt #6 Aug 13 07:20:59.003266 ignition[942]: GET result: OK Aug 13 07:21:00.060590 ignition[942]: Ignition finished successfully Aug 13 07:21:00.065609 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 07:21:00.097315 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 07:21:00.103579 ignition[962]: Ignition 2.19.0 Aug 13 07:21:00.103584 ignition[962]: Stage: disks Aug 13 07:21:00.103700 ignition[962]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:00.103707 ignition[962]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:00.104305 ignition[962]: disks: disks passed Aug 13 07:21:00.104308 ignition[962]: POST message to Packet Timeline Aug 13 07:21:00.104319 ignition[962]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:01.525110 ignition[962]: GET result: OK Aug 13 07:21:02.415724 ignition[962]: Ignition finished successfully Aug 13 07:21:02.418694 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 07:21:02.434341 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 07:21:02.452346 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 07:21:02.473363 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:21:02.495373 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:21:02.516382 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:21:02.550342 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 07:21:02.585397 systemd-fsck[980]: ROOT: clean, 14/553520 files, 52654/553472 blocks Aug 13 07:21:02.595517 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 07:21:02.626484 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 07:21:02.724755 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 07:21:02.739304 kernel: EXT4-fs (sdb9): mounted filesystem 98cc0201-e9ec-4d2c-8a62-5b521bf9317d r/w with ordered data mode. Quota mode: none. Aug 13 07:21:02.725018 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 07:21:02.757273 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:21:02.760857 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 07:21:02.776160 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Aug 13 07:21:02.806069 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sdb6 scanned by mount (989) Aug 13 07:21:02.806085 kernel: BTRFS info (device sdb6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:21:02.838011 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:21:02.842552 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Aug 13 07:21:02.907049 kernel: BTRFS info (device sdb6): using free space tree Aug 13 07:21:02.907071 kernel: BTRFS info (device sdb6): enabling ssd optimizations Aug 13 07:21:02.907085 kernel: BTRFS info (device sdb6): auto enabling async discard Aug 13 07:21:02.907045 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 07:21:02.907072 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:21:02.955619 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:21:02.979327 coreos-metadata[991]: Aug 13 07:21:02.966 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 07:21:02.964355 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 07:21:03.002318 coreos-metadata[992]: Aug 13 07:21:02.966 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 07:21:03.001255 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 07:21:03.040243 initrd-setup-root[1021]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 07:21:03.050179 initrd-setup-root[1028]: cut: /sysroot/etc/group: No such file or directory Aug 13 07:21:03.060181 initrd-setup-root[1035]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 07:21:03.071193 initrd-setup-root[1042]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 07:21:03.087033 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 07:21:03.111339 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 07:21:03.147192 kernel: BTRFS info (device sdb6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:21:03.129806 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 07:21:03.155844 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 07:21:03.177838 ignition[1109]: INFO : Ignition 2.19.0 Aug 13 07:21:03.177838 ignition[1109]: INFO : Stage: mount Aug 13 07:21:03.186178 ignition[1109]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:03.186178 ignition[1109]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:03.186178 ignition[1109]: INFO : mount: mount passed Aug 13 07:21:03.186178 ignition[1109]: INFO : POST message to Packet Timeline Aug 13 07:21:03.186178 ignition[1109]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:03.182829 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 07:21:04.005176 coreos-metadata[991]: Aug 13 07:21:04.005 INFO Fetch successful Aug 13 07:21:04.015161 coreos-metadata[992]: Aug 13 07:21:04.009 INFO Fetch successful Aug 13 07:21:04.036520 coreos-metadata[991]: Aug 13 07:21:04.036 INFO wrote hostname ci-4081.3.5-a-3ca3f1276f to /sysroot/etc/hostname Aug 13 07:21:04.049448 ignition[1109]: INFO : GET result: OK Aug 13 07:21:04.037796 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 07:21:04.060997 systemd[1]: flatcar-static-network.service: Deactivated successfully. Aug 13 07:21:04.061270 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Aug 13 07:21:04.502204 ignition[1109]: INFO : Ignition finished successfully Aug 13 07:21:04.506052 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 07:21:04.533295 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 07:21:04.544108 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:21:04.614827 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sdb6 scanned by mount (1136) Aug 13 07:21:04.614854 kernel: BTRFS info (device sdb6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:21:04.634927 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:21:04.652609 kernel: BTRFS info (device sdb6): using free space tree Aug 13 07:21:04.689968 kernel: BTRFS info (device sdb6): enabling ssd optimizations Aug 13 07:21:04.689984 kernel: BTRFS info (device sdb6): auto enabling async discard Aug 13 07:21:04.702677 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:21:04.730821 ignition[1153]: INFO : Ignition 2.19.0 Aug 13 07:21:04.730821 ignition[1153]: INFO : Stage: files Aug 13 07:21:04.745309 ignition[1153]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:04.745309 ignition[1153]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:04.745309 ignition[1153]: DEBUG : files: compiled without relabeling support, skipping Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 07:21:04.745309 ignition[1153]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Aug 13 07:21:04.745309 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Aug 13 07:21:04.735015 unknown[1153]: wrote ssh authorized keys file for user: core Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:04.913265 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:05.163421 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Aug 13 07:21:05.644359 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Aug 13 07:21:05.845251 ignition[1153]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 07:21:05.845251 ignition[1153]: INFO : files: op(c): [started] processing unit "containerd.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(c): [finished] processing unit "containerd.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:21:05.874367 ignition[1153]: INFO : files: files passed Aug 13 07:21:05.874367 ignition[1153]: INFO : POST message to Packet Timeline Aug 13 07:21:05.874367 ignition[1153]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:06.727274 ignition[1153]: INFO : GET result: OK Aug 13 07:21:07.145826 ignition[1153]: INFO : Ignition finished successfully Aug 13 07:21:07.148676 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 07:21:07.187384 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 07:21:07.198801 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 07:21:07.220602 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 07:21:07.220708 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 07:21:07.260907 initrd-setup-root-after-ignition[1191]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:21:07.260907 initrd-setup-root-after-ignition[1191]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:21:07.299380 initrd-setup-root-after-ignition[1195]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:21:07.265371 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:21:07.276341 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 07:21:07.326463 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 07:21:07.410248 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 07:21:07.410300 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 07:21:07.428449 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 07:21:07.439408 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 07:21:07.466440 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 07:21:07.487289 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 07:21:07.530935 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:21:07.561623 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 07:21:07.580658 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:21:07.584261 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:21:07.616496 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 07:21:07.635420 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 07:21:07.635593 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:21:07.663810 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 07:21:07.684704 systemd[1]: Stopped target basic.target - Basic System. Aug 13 07:21:07.703694 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 07:21:07.722794 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:21:07.743680 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 07:21:07.764672 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 07:21:07.785427 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:21:07.806706 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 07:21:07.827773 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 07:21:07.847650 systemd[1]: Stopped target swap.target - Swaps. Aug 13 07:21:07.865662 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 07:21:07.866088 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:21:07.900556 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:21:07.910703 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:21:07.932549 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 07:21:07.932910 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:21:07.954570 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 07:21:07.954975 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 07:21:07.987662 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 07:21:07.988154 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:21:08.008890 systemd[1]: Stopped target paths.target - Path Units. Aug 13 07:21:08.027545 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 07:21:08.027969 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:21:08.048700 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 07:21:08.066653 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 07:21:08.084665 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 07:21:08.084973 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:21:08.104738 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 07:21:08.105044 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:21:08.127777 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 07:21:08.128210 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:21:08.148770 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 07:21:08.265289 ignition[1215]: INFO : Ignition 2.19.0 Aug 13 07:21:08.265289 ignition[1215]: INFO : Stage: umount Aug 13 07:21:08.265289 ignition[1215]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:21:08.265289 ignition[1215]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 07:21:08.265289 ignition[1215]: INFO : umount: umount passed Aug 13 07:21:08.265289 ignition[1215]: INFO : POST message to Packet Timeline Aug 13 07:21:08.265289 ignition[1215]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 07:21:08.149181 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 07:21:08.166783 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Aug 13 07:21:08.167195 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 07:21:08.196241 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 07:21:08.211360 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 07:21:08.211489 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:21:08.249268 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 07:21:08.250406 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 07:21:08.250488 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:21:08.276331 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 07:21:08.276411 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:21:08.326344 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 07:21:08.328002 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 07:21:08.328279 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 07:21:08.341940 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 07:21:08.342201 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 07:21:09.150216 ignition[1215]: INFO : GET result: OK Aug 13 07:21:09.595137 ignition[1215]: INFO : Ignition finished successfully Aug 13 07:21:09.598209 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 07:21:09.598515 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 07:21:09.615396 systemd[1]: Stopped target network.target - Network. Aug 13 07:21:09.631303 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 07:21:09.631486 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 07:21:09.649488 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 07:21:09.649654 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 07:21:09.669402 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 07:21:09.669544 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 07:21:09.688475 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 07:21:09.688647 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 07:21:09.707471 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 07:21:09.707643 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 07:21:09.726878 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 07:21:09.737203 systemd-networkd[922]: enp1s0f1np1: DHCPv6 lease lost Aug 13 07:21:09.744296 systemd-networkd[922]: enp1s0f0np0: DHCPv6 lease lost Aug 13 07:21:09.744585 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 07:21:09.765210 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 07:21:09.765499 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 07:21:09.785444 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 07:21:09.785790 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 07:21:09.806191 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 07:21:09.806378 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:21:09.837319 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 07:21:09.844441 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 07:21:09.844599 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:21:09.872412 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 07:21:09.872560 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:21:09.890362 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 07:21:09.890503 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 07:21:09.908471 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 07:21:09.908640 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:21:09.929799 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:21:09.966846 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 07:21:09.967271 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:21:09.971672 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 07:21:09.971829 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 07:21:09.999423 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 07:21:09.999537 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:21:10.017335 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 07:21:10.017479 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:21:10.047627 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 07:21:10.047798 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 07:21:10.078496 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:21:10.078648 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:21:10.139261 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 07:21:10.168225 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 07:21:10.168270 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:21:10.186224 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:21:10.422260 systemd-journald[267]: Received SIGTERM from PID 1 (systemd). Aug 13 07:21:10.186281 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:21:10.210455 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 07:21:10.210712 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 07:21:10.271623 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 07:21:10.271922 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 07:21:10.291085 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 07:21:10.330458 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 07:21:10.350722 systemd[1]: Switching root. Aug 13 07:21:10.495286 systemd-journald[267]: Journal stopped Aug 13 07:21:13.016596 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 07:21:13.016615 kernel: SELinux: policy capability open_perms=1 Aug 13 07:21:13.016694 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 07:21:13.016702 kernel: SELinux: policy capability always_check_network=0 Aug 13 07:21:13.016709 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 07:21:13.016715 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 07:21:13.016722 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 07:21:13.016728 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 07:21:13.016735 kernel: audit: type=1403 audit(1755069670.775:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Aug 13 07:21:13.016743 systemd[1]: Successfully loaded SELinux policy in 154.695ms. Aug 13 07:21:13.016753 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 7.261ms. Aug 13 07:21:13.016761 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 07:21:13.016768 systemd[1]: Detected architecture x86-64. Aug 13 07:21:13.016776 systemd[1]: Detected first boot. Aug 13 07:21:13.016783 systemd[1]: Hostname set to . Aug 13 07:21:13.016793 systemd[1]: Initializing machine ID from random generator. Aug 13 07:21:13.016800 zram_generator::config[1283]: No configuration found. Aug 13 07:21:13.016808 systemd[1]: Populated /etc with preset unit settings. Aug 13 07:21:13.016816 systemd[1]: Queued start job for default target multi-user.target. Aug 13 07:21:13.016823 systemd[1]: Unnecessary job was removed for dev-sdb6.device - /dev/sdb6. Aug 13 07:21:13.016831 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Aug 13 07:21:13.016839 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Aug 13 07:21:13.016847 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Aug 13 07:21:13.016855 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Aug 13 07:21:13.016863 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Aug 13 07:21:13.016871 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Aug 13 07:21:13.016878 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Aug 13 07:21:13.016886 systemd[1]: Created slice user.slice - User and Session Slice. Aug 13 07:21:13.016894 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:21:13.016903 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:21:13.016911 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Aug 13 07:21:13.016918 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Aug 13 07:21:13.016926 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Aug 13 07:21:13.016935 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:21:13.016943 systemd[1]: Expecting device dev-ttyS1.device - /dev/ttyS1... Aug 13 07:21:13.016951 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:21:13.016958 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Aug 13 07:21:13.016967 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:21:13.016975 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:21:13.016983 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:21:13.016993 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:21:13.017001 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Aug 13 07:21:13.017009 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Aug 13 07:21:13.017017 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 07:21:13.017026 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Aug 13 07:21:13.017034 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:21:13.017042 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:21:13.017050 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:21:13.017058 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Aug 13 07:21:13.017069 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Aug 13 07:21:13.017080 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Aug 13 07:21:13.017088 systemd[1]: Mounting media.mount - External Media Directory... Aug 13 07:21:13.017097 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:13.017105 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Aug 13 07:21:13.017112 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Aug 13 07:21:13.017120 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Aug 13 07:21:13.017129 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Aug 13 07:21:13.017138 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:21:13.017146 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:21:13.017154 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Aug 13 07:21:13.017163 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:21:13.017171 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:21:13.017179 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:21:13.017187 kernel: ACPI: bus type drm_connector registered Aug 13 07:21:13.017194 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Aug 13 07:21:13.017202 kernel: fuse: init (API version 7.39) Aug 13 07:21:13.017211 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:21:13.017219 kernel: loop: module loaded Aug 13 07:21:13.017226 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 07:21:13.017235 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Aug 13 07:21:13.017243 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Aug 13 07:21:13.017251 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:21:13.017259 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:21:13.017277 systemd-journald[1406]: Collecting audit messages is disabled. Aug 13 07:21:13.017296 systemd-journald[1406]: Journal started Aug 13 07:21:13.017314 systemd-journald[1406]: Runtime Journal (/run/log/journal/a5a435dea8de4ad88be1dabcbf871574) is 8.0M, max 639.9M, 631.9M free. Aug 13 07:21:13.048106 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Aug 13 07:21:13.083069 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Aug 13 07:21:13.118117 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:21:13.169073 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:13.191267 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:21:13.200953 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Aug 13 07:21:13.211333 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Aug 13 07:21:13.221336 systemd[1]: Mounted media.mount - External Media Directory. Aug 13 07:21:13.231316 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Aug 13 07:21:13.241276 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Aug 13 07:21:13.251309 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Aug 13 07:21:13.261507 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Aug 13 07:21:13.272548 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:21:13.283609 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 07:21:13.283846 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Aug 13 07:21:13.296270 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:21:13.296778 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:21:13.309034 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:21:13.309518 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:21:13.320031 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:21:13.320515 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:21:13.332035 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 07:21:13.332508 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Aug 13 07:21:13.343029 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:21:13.343541 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:21:13.354333 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:21:13.365180 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 07:21:13.378133 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Aug 13 07:21:13.391101 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:21:13.411335 systemd[1]: Reached target network-pre.target - Preparation for Network. Aug 13 07:21:13.433219 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Aug 13 07:21:13.444036 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Aug 13 07:21:13.454241 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 07:21:13.455844 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Aug 13 07:21:13.466844 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Aug 13 07:21:13.478206 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:21:13.485538 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Aug 13 07:21:13.488105 systemd-journald[1406]: Time spent on flushing to /var/log/journal/a5a435dea8de4ad88be1dabcbf871574 is 12.690ms for 1355 entries. Aug 13 07:21:13.488105 systemd-journald[1406]: System Journal (/var/log/journal/a5a435dea8de4ad88be1dabcbf871574) is 8.0M, max 195.6M, 187.6M free. Aug 13 07:21:13.526266 systemd-journald[1406]: Received client request to flush runtime journal. Aug 13 07:21:13.504206 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:21:13.504863 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:21:13.510623 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 07:21:13.531053 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Aug 13 07:21:13.543430 systemd-tmpfiles[1443]: ACLs are not supported, ignoring. Aug 13 07:21:13.543440 systemd-tmpfiles[1443]: ACLs are not supported, ignoring. Aug 13 07:21:13.549507 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Aug 13 07:21:13.560289 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Aug 13 07:21:13.571381 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Aug 13 07:21:13.582337 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Aug 13 07:21:13.593331 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:21:13.603309 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 07:21:13.616738 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Aug 13 07:21:13.640233 systemd[1]: Starting systemd-sysusers.service - Create System Users... Aug 13 07:21:13.650435 udevadm[1450]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Aug 13 07:21:13.657455 systemd[1]: Finished systemd-sysusers.service - Create System Users. Aug 13 07:21:13.678255 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:21:13.685648 systemd-tmpfiles[1464]: ACLs are not supported, ignoring. Aug 13 07:21:13.685659 systemd-tmpfiles[1464]: ACLs are not supported, ignoring. Aug 13 07:21:13.689441 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:21:13.838242 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Aug 13 07:21:13.860334 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:21:13.873096 systemd-udevd[1472]: Using default interface naming scheme 'v255'. Aug 13 07:21:13.887603 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:21:13.904264 systemd[1]: Found device dev-ttyS1.device - /dev/ttyS1. Aug 13 07:21:13.922015 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input2 Aug 13 07:21:13.922062 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 37 scanned by (udev-worker) (1522) Aug 13 07:21:13.922103 kernel: ACPI: button: Sleep Button [SLPB] Aug 13 07:21:13.989119 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Aug 13 07:21:13.995071 kernel: IPMI message handler: version 39.2 Aug 13 07:21:13.995090 kernel: ACPI: button: Power Button [PWRF] Aug 13 07:21:14.020960 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Aug 13 07:21:14.026075 kernel: mousedev: PS/2 mouse device common for all mice Aug 13 07:21:14.070080 kernel: ipmi device interface Aug 13 07:21:14.093080 kernel: i801_smbus 0000:00:1f.4: SPD Write Disable is set Aug 13 07:21:14.093407 kernel: i801_smbus 0000:00:1f.4: SMBus using PCI interrupt Aug 13 07:21:14.115253 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:21:14.133070 kernel: i2c i2c-0: 2/4 memory slots populated (from DMI) Aug 13 07:21:14.142068 kernel: ipmi_si: IPMI System Interface driver Aug 13 07:21:14.142088 kernel: mei_me 0000:00:16.0: Device doesn't have valid ME Interface Aug 13 07:21:14.142182 kernel: mei_me 0000:00:16.4: Device doesn't have valid ME Interface Aug 13 07:21:14.153390 kernel: ipmi_si dmi-ipmi-si.0: ipmi_platform: probing via SMBIOS Aug 13 07:21:14.154079 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:21:14.227587 kernel: ipmi_platform: ipmi_si: SMBIOS: io 0xca2 regsize 1 spacing 1 irq 0 Aug 13 07:21:14.227607 kernel: ipmi_si: Adding SMBIOS-specified kcs state machine Aug 13 07:21:14.227616 kernel: ipmi_si IPI0001:00: ipmi_platform: probing via ACPI Aug 13 07:21:14.227704 kernel: ipmi_si IPI0001:00: ipmi_platform: [io 0x0ca2] regsize 1 spacing 1 irq 0 Aug 13 07:21:14.300069 kernel: ipmi_si dmi-ipmi-si.0: Removing SMBIOS-specified kcs state machine in favor of ACPI Aug 13 07:21:14.300305 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Aug 13 07:21:14.322124 kernel: ipmi_si: Adding ACPI-specified kcs state machine Aug 13 07:21:14.322154 kernel: ipmi_si: Trying ACPI-specified kcs state machine at i/o address 0xca2, slave address 0x20, irq 0 Aug 13 07:21:14.375071 kernel: iTCO_vendor_support: vendor-support=0 Aug 13 07:21:14.375107 kernel: ipmi_si IPI0001:00: The BMC does not support clearing the recv irq bit, compensating, but the BMC needs to be fixed. Aug 13 07:21:14.464122 kernel: iTCO_wdt iTCO_wdt: Found a Intel PCH TCO device (Version=6, TCOBASE=0x0400) Aug 13 07:21:14.464269 kernel: ipmi_si IPI0001:00: IPMI message handler: Found new BMC (man_id: 0x002a7c, prod_id: 0x1b0f, dev_id: 0x20) Aug 13 07:21:14.464359 kernel: iTCO_wdt iTCO_wdt: initialized. heartbeat=30 sec (nowayout=0) Aug 13 07:21:14.499522 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:21:14.518391 systemd[1]: Started systemd-userdbd.service - User Database Manager. Aug 13 07:21:14.531070 kernel: intel_rapl_common: Found RAPL domain package Aug 13 07:21:14.531098 kernel: intel_rapl_common: Found RAPL domain core Aug 13 07:21:14.531108 kernel: intel_rapl_common: Found RAPL domain dram Aug 13 07:21:14.586073 kernel: ipmi_si IPI0001:00: IPMI kcs interface initialized Aug 13 07:21:14.606113 kernel: ipmi_ssif: IPMI SSIF Interface driver Aug 13 07:21:14.606918 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Aug 13 07:21:14.628582 systemd-networkd[1561]: lo: Link UP Aug 13 07:21:14.628585 systemd-networkd[1561]: lo: Gained carrier Aug 13 07:21:14.631326 systemd-networkd[1561]: bond0: netdev ready Aug 13 07:21:14.632263 systemd-networkd[1561]: Enumeration completed Aug 13 07:21:14.633151 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Aug 13 07:21:14.640654 lvm[1587]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:21:14.640950 systemd-networkd[1561]: enp1s0f0np0: Configuring with /etc/systemd/network/10-1c:34:da:42:c5:9c.network. Aug 13 07:21:14.645153 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:21:14.656579 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Aug 13 07:21:14.682556 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Aug 13 07:21:14.694458 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:21:14.712124 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Aug 13 07:21:14.713909 lvm[1595]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:21:14.760789 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Aug 13 07:21:14.772283 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 07:21:14.783111 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 07:21:14.783129 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:21:14.792110 systemd[1]: Reached target machines.target - Containers. Aug 13 07:21:14.801709 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Aug 13 07:21:14.825117 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Aug 13 07:21:14.837678 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Aug 13 07:21:14.848148 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:21:14.848609 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Aug 13 07:21:14.866457 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Aug 13 07:21:14.877865 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Aug 13 07:21:14.878337 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Aug 13 07:21:14.904165 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Aug 13 07:21:14.908068 kernel: loop0: detected capacity change from 0 to 140768 Aug 13 07:21:14.920091 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 07:21:14.920508 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Aug 13 07:21:14.951073 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 07:21:15.008071 kernel: loop1: detected capacity change from 0 to 8 Aug 13 07:21:15.051069 kernel: loop2: detected capacity change from 0 to 221472 Aug 13 07:21:15.122069 kernel: loop3: detected capacity change from 0 to 142488 Aug 13 07:21:15.178398 ldconfig[1601]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 07:21:15.179928 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Aug 13 07:21:15.183067 kernel: loop4: detected capacity change from 0 to 140768 Aug 13 07:21:15.217556 kernel: loop5: detected capacity change from 0 to 8 Aug 13 07:21:15.217580 kernel: loop6: detected capacity change from 0 to 221472 Aug 13 07:21:15.255068 kernel: loop7: detected capacity change from 0 to 142488 Aug 13 07:21:15.264952 (sd-merge)[1617]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-packet'. Aug 13 07:21:15.265191 (sd-merge)[1617]: Merged extensions into '/usr'. Aug 13 07:21:15.267225 systemd[1]: Reloading requested from client PID 1605 ('systemd-sysext') (unit systemd-sysext.service)... Aug 13 07:21:15.267231 systemd[1]: Reloading... Aug 13 07:21:15.302077 zram_generator::config[1646]: No configuration found. Aug 13 07:21:15.363023 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:21:15.414653 systemd[1]: Reloading finished in 147 ms. Aug 13 07:21:15.426303 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Aug 13 07:21:15.446183 systemd[1]: Starting ensure-sysext.service... Aug 13 07:21:15.453774 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:21:15.471136 systemd[1]: Reloading requested from client PID 1707 ('systemctl') (unit ensure-sysext.service)... Aug 13 07:21:15.471143 systemd[1]: Reloading... Aug 13 07:21:15.478807 systemd-tmpfiles[1708]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 07:21:15.479011 systemd-tmpfiles[1708]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Aug 13 07:21:15.479512 systemd-tmpfiles[1708]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 07:21:15.479679 systemd-tmpfiles[1708]: ACLs are not supported, ignoring. Aug 13 07:21:15.479716 systemd-tmpfiles[1708]: ACLs are not supported, ignoring. Aug 13 07:21:15.482478 systemd-tmpfiles[1708]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:21:15.482482 systemd-tmpfiles[1708]: Skipping /boot Aug 13 07:21:15.488155 systemd-tmpfiles[1708]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:21:15.488161 systemd-tmpfiles[1708]: Skipping /boot Aug 13 07:21:15.503071 zram_generator::config[1736]: No configuration found. Aug 13 07:21:15.567013 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:21:15.619251 systemd[1]: Reloading finished in 147 ms. Aug 13 07:21:15.629114 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Aug 13 07:21:15.651131 kernel: bond0: (slave enp1s0f0np0): Enslaving as a backup interface with an up link Aug 13 07:21:15.652221 systemd-networkd[1561]: enp1s0f1np1: Configuring with /etc/systemd/network/10-1c:34:da:42:c5:9d.network. Aug 13 07:21:15.656811 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:21:15.678886 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 07:21:15.689493 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Aug 13 07:21:15.700909 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Aug 13 07:21:15.707441 augenrules[1818]: No rules Aug 13 07:21:15.713197 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:21:15.724751 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Aug 13 07:21:15.736474 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 07:21:15.747518 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Aug 13 07:21:15.758445 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Aug 13 07:21:15.779691 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:15.779824 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:21:15.780540 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:21:15.789852 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:21:15.802774 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:21:15.813128 systemd-resolved[1824]: Positive Trust Anchors: Aug 13 07:21:15.813136 systemd-resolved[1824]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:21:15.813153 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:21:15.813160 systemd-resolved[1824]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:21:15.814013 systemd[1]: Starting systemd-update-done.service - Update is Completed... Aug 13 07:21:15.816373 systemd-resolved[1824]: Using system hostname 'ci-4081.3.5-a-3ca3f1276f'. Aug 13 07:21:15.825174 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 07:21:15.825239 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:15.825996 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Aug 13 07:21:15.843145 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Aug 13 07:21:15.843439 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:21:15.843529 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:21:15.862401 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:21:15.862484 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:21:15.864286 systemd-networkd[1561]: bond0: Configuring with /etc/systemd/network/05-bond0.network. Aug 13 07:21:15.865139 kernel: bond0: (slave enp1s0f1np1): Enslaving as a backup interface with an up link Aug 13 07:21:15.865425 systemd-networkd[1561]: enp1s0f0np0: Link UP Aug 13 07:21:15.865577 systemd-networkd[1561]: enp1s0f0np0: Gained carrier Aug 13 07:21:15.885123 kernel: bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond Aug 13 07:21:15.891266 systemd-networkd[1561]: enp1s0f1np1: Reconfiguring with /etc/systemd/network/10-1c:34:da:42:c5:9c.network. Aug 13 07:21:15.891410 systemd-networkd[1561]: enp1s0f1np1: Link UP Aug 13 07:21:15.891558 systemd-networkd[1561]: enp1s0f1np1: Gained carrier Aug 13 07:21:15.895335 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:21:15.902239 systemd-networkd[1561]: bond0: Link UP Aug 13 07:21:15.902399 systemd-networkd[1561]: bond0: Gained carrier Aug 13 07:21:15.905436 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:21:15.905519 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:21:15.915430 systemd[1]: Finished systemd-update-done.service - Update is Completed. Aug 13 07:21:15.927729 systemd[1]: Reached target network.target - Network. Aug 13 07:21:15.936227 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:21:15.947223 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:15.947344 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:21:15.959262 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:21:15.969929 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:21:15.990069 kernel: bond0: (slave enp1s0f0np0): link status definitely up, 10000 Mbps full duplex Aug 13 07:21:15.990089 kernel: bond0: active interface up! Aug 13 07:21:16.009246 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:21:16.019185 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:21:16.019261 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 07:21:16.019310 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:16.019943 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:21:16.020026 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:21:16.031373 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:21:16.031454 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:21:16.042346 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:21:16.042426 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:21:16.054280 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:16.054414 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:21:16.063271 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:21:16.073730 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:21:16.089308 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:21:16.108752 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:21:16.111139 kernel: bond0: (slave enp1s0f1np1): link status definitely up, 10000 Mbps full duplex Aug 13 07:21:16.120226 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:21:16.120315 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 07:21:16.120379 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:21:16.121045 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:21:16.121183 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:21:16.132612 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:21:16.132689 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:21:16.142348 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:21:16.142426 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:21:16.153340 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:21:16.153417 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:21:16.164117 systemd[1]: Finished ensure-sysext.service. Aug 13 07:21:16.173609 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:21:16.173640 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:21:16.184208 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Aug 13 07:21:16.222781 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Aug 13 07:21:16.234270 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:21:16.244197 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Aug 13 07:21:16.255152 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Aug 13 07:21:16.266147 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Aug 13 07:21:16.277147 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 07:21:16.277165 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:21:16.285140 systemd[1]: Reached target time-set.target - System Time Set. Aug 13 07:21:16.295213 systemd[1]: Started logrotate.timer - Daily rotation of log files. Aug 13 07:21:16.305196 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Aug 13 07:21:16.316140 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:21:16.324706 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Aug 13 07:21:16.334923 systemd[1]: Starting docker.socket - Docker Socket for the API... Aug 13 07:21:16.343919 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Aug 13 07:21:16.353384 systemd[1]: Listening on docker.socket - Docker Socket for the API. Aug 13 07:21:16.363173 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:21:16.373153 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:21:16.381222 systemd[1]: System is tainted: cgroupsv1 Aug 13 07:21:16.381244 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:21:16.381259 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:21:16.392145 systemd[1]: Starting containerd.service - containerd container runtime... Aug 13 07:21:16.402914 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Aug 13 07:21:16.420439 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Aug 13 07:21:16.428788 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Aug 13 07:21:16.433156 coreos-metadata[1878]: Aug 13 07:21:16.433 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 07:21:16.439513 dbus-daemon[1879]: [system] SELinux support is enabled Aug 13 07:21:16.439926 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Aug 13 07:21:16.441662 jq[1882]: false Aug 13 07:21:16.449204 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Aug 13 07:21:16.449862 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Aug 13 07:21:16.457187 extend-filesystems[1884]: Found loop4 Aug 13 07:21:16.457187 extend-filesystems[1884]: Found loop5 Aug 13 07:21:16.457187 extend-filesystems[1884]: Found loop6 Aug 13 07:21:16.525160 kernel: EXT4-fs (sdb9): resizing filesystem from 553472 to 116605649 blocks Aug 13 07:21:16.525179 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 37 scanned by (udev-worker) (1481) Aug 13 07:21:16.459825 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Aug 13 07:21:16.525259 extend-filesystems[1884]: Found loop7 Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sda Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb1 Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb2 Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb3 Aug 13 07:21:16.525259 extend-filesystems[1884]: Found usr Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb4 Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb6 Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb7 Aug 13 07:21:16.525259 extend-filesystems[1884]: Found sdb9 Aug 13 07:21:16.525259 extend-filesystems[1884]: Checking size of /dev/sdb9 Aug 13 07:21:16.525259 extend-filesystems[1884]: Resized partition /dev/sdb9 Aug 13 07:21:16.656306 extend-filesystems[1893]: resize2fs 1.47.1 (20-May-2024) Aug 13 07:21:16.532678 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Aug 13 07:21:16.564483 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Aug 13 07:21:16.583371 systemd[1]: Starting systemd-logind.service - User Login Management... Aug 13 07:21:16.611145 systemd[1]: Starting tcsd.service - TCG Core Services Daemon... Aug 13 07:21:16.625843 systemd[1]: Starting update-engine.service - Update Engine... Aug 13 07:21:16.635856 systemd-logind[1908]: Watching system buttons on /dev/input/event3 (Power Button) Aug 13 07:21:16.667763 update_engine[1913]: I20250813 07:21:16.657246 1913 main.cc:92] Flatcar Update Engine starting Aug 13 07:21:16.667763 update_engine[1913]: I20250813 07:21:16.657946 1913 update_check_scheduler.cc:74] Next update check in 2m38s Aug 13 07:21:16.635866 systemd-logind[1908]: Watching system buttons on /dev/input/event2 (Sleep Button) Aug 13 07:21:16.635876 systemd-logind[1908]: Watching system buttons on /dev/input/event0 (HID 0557:2419) Aug 13 07:21:16.636011 systemd-logind[1908]: New seat seat0. Aug 13 07:21:16.650098 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Aug 13 07:21:16.656431 systemd[1]: Started dbus.service - D-Bus System Message Bus. Aug 13 07:21:16.668986 jq[1914]: true Aug 13 07:21:16.683396 systemd[1]: Started systemd-logind.service - User Login Management. Aug 13 07:21:16.701198 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 07:21:16.701327 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Aug 13 07:21:16.701509 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 07:21:16.701623 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Aug 13 07:21:16.711515 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 07:21:16.711637 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Aug 13 07:21:16.732991 jq[1919]: true Aug 13 07:21:16.733752 (ntainerd)[1920]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Aug 13 07:21:16.736893 dbus-daemon[1879]: [system] Successfully activated service 'org.freedesktop.systemd1' Aug 13 07:21:16.739431 tar[1918]: linux-amd64/helm Aug 13 07:21:16.744175 systemd[1]: tcsd.service: Skipped due to 'exec-condition'. Aug 13 07:21:16.744336 systemd[1]: Condition check resulted in tcsd.service - TCG Core Services Daemon being skipped. Aug 13 07:21:16.746088 systemd[1]: Started update-engine.service - Update Engine. Aug 13 07:21:16.760382 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 07:21:16.760486 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Aug 13 07:21:16.771200 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 07:21:16.771294 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Aug 13 07:21:16.782495 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Aug 13 07:21:16.788088 bash[1948]: Updated "/home/core/.ssh/authorized_keys" Aug 13 07:21:16.797290 systemd[1]: Started locksmithd.service - Cluster reboot manager. Aug 13 07:21:16.797497 sshd_keygen[1911]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 07:21:16.809153 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Aug 13 07:21:16.818470 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Aug 13 07:21:16.822262 locksmithd[1951]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 07:21:16.841244 systemd[1]: Starting issuegen.service - Generate /run/issue... Aug 13 07:21:16.850536 systemd[1]: Starting sshkeys.service... Aug 13 07:21:16.857444 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 07:21:16.857582 systemd[1]: Finished issuegen.service - Generate /run/issue. Aug 13 07:21:16.871769 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Aug 13 07:21:16.894484 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Aug 13 07:21:16.905992 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Aug 13 07:21:16.916900 containerd[1920]: time="2025-08-13T07:21:16.916855019Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Aug 13 07:21:16.918538 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Aug 13 07:21:16.930993 systemd[1]: Started getty@tty1.service - Getty on tty1. Aug 13 07:21:16.931706 containerd[1920]: time="2025-08-13T07:21:16.931636029Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932248 coreos-metadata[1984]: Aug 13 07:21:16.932 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 07:21:16.932413 containerd[1920]: time="2025-08-13T07:21:16.932399568Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.100-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932439 containerd[1920]: time="2025-08-13T07:21:16.932415438Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Aug 13 07:21:16.932439 containerd[1920]: time="2025-08-13T07:21:16.932424781Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Aug 13 07:21:16.932563 containerd[1920]: time="2025-08-13T07:21:16.932525982Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Aug 13 07:21:16.932563 containerd[1920]: time="2025-08-13T07:21:16.932542914Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932603 containerd[1920]: time="2025-08-13T07:21:16.932577492Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932603 containerd[1920]: time="2025-08-13T07:21:16.932591989Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932761 containerd[1920]: time="2025-08-13T07:21:16.932723156Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932761 containerd[1920]: time="2025-08-13T07:21:16.932733562Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932761 containerd[1920]: time="2025-08-13T07:21:16.932746906Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932761 containerd[1920]: time="2025-08-13T07:21:16.932753213Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932832 containerd[1920]: time="2025-08-13T07:21:16.932806308Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:21:16.932966 containerd[1920]: time="2025-08-13T07:21:16.932934099Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:21:16.933026 containerd[1920]: time="2025-08-13T07:21:16.933017875Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:21:16.933042 containerd[1920]: time="2025-08-13T07:21:16.933027152Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Aug 13 07:21:16.933093 containerd[1920]: time="2025-08-13T07:21:16.933085754Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Aug 13 07:21:16.933122 containerd[1920]: time="2025-08-13T07:21:16.933116102Z" level=info msg="metadata content store policy set" policy=shared Aug 13 07:21:16.939946 systemd[1]: Started serial-getty@ttyS1.service - Serial Getty on ttyS1. Aug 13 07:21:16.944405 containerd[1920]: time="2025-08-13T07:21:16.944391879Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Aug 13 07:21:16.944440 containerd[1920]: time="2025-08-13T07:21:16.944418122Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Aug 13 07:21:16.944459 containerd[1920]: time="2025-08-13T07:21:16.944434799Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Aug 13 07:21:16.944484 containerd[1920]: time="2025-08-13T07:21:16.944474338Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Aug 13 07:21:16.944500 containerd[1920]: time="2025-08-13T07:21:16.944487024Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Aug 13 07:21:16.944578 containerd[1920]: time="2025-08-13T07:21:16.944566578Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Aug 13 07:21:16.944750 containerd[1920]: time="2025-08-13T07:21:16.944741915Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Aug 13 07:21:16.944809 containerd[1920]: time="2025-08-13T07:21:16.944800788Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Aug 13 07:21:16.944827 containerd[1920]: time="2025-08-13T07:21:16.944812552Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Aug 13 07:21:16.944842 containerd[1920]: time="2025-08-13T07:21:16.944826716Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Aug 13 07:21:16.944861 containerd[1920]: time="2025-08-13T07:21:16.944842647Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944861 containerd[1920]: time="2025-08-13T07:21:16.944852348Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944861 containerd[1920]: time="2025-08-13T07:21:16.944859473Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944904 containerd[1920]: time="2025-08-13T07:21:16.944867259Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944904 containerd[1920]: time="2025-08-13T07:21:16.944874770Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944904 containerd[1920]: time="2025-08-13T07:21:16.944882655Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944904 containerd[1920]: time="2025-08-13T07:21:16.944889757Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944904 containerd[1920]: time="2025-08-13T07:21:16.944895647Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944907235Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944914801Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944921648Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944929387Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944936757Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944944095Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944953663Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944961285Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.944973 containerd[1920]: time="2025-08-13T07:21:16.944967912Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.944975812Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.944981997Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.944988166Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.944994788Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945003010Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945014388Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945023999Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945035130Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945067433Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945079659Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945086061Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945092564Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Aug 13 07:21:16.945108 containerd[1920]: time="2025-08-13T07:21:16.945098274Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945292 containerd[1920]: time="2025-08-13T07:21:16.945105071Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Aug 13 07:21:16.945292 containerd[1920]: time="2025-08-13T07:21:16.945113448Z" level=info msg="NRI interface is disabled by configuration." Aug 13 07:21:16.945292 containerd[1920]: time="2025-08-13T07:21:16.945119474Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Aug 13 07:21:16.945338 containerd[1920]: time="2025-08-13T07:21:16.945292070Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Aug 13 07:21:16.945338 containerd[1920]: time="2025-08-13T07:21:16.945326436Z" level=info msg="Connect containerd service" Aug 13 07:21:16.945429 containerd[1920]: time="2025-08-13T07:21:16.945347498Z" level=info msg="using legacy CRI server" Aug 13 07:21:16.945429 containerd[1920]: time="2025-08-13T07:21:16.945352362Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Aug 13 07:21:16.945429 containerd[1920]: time="2025-08-13T07:21:16.945400038Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Aug 13 07:21:16.945692 containerd[1920]: time="2025-08-13T07:21:16.945681641Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 07:21:16.945790 containerd[1920]: time="2025-08-13T07:21:16.945773552Z" level=info msg="Start subscribing containerd event" Aug 13 07:21:16.945812 containerd[1920]: time="2025-08-13T07:21:16.945801051Z" level=info msg="Start recovering state" Aug 13 07:21:16.945852 containerd[1920]: time="2025-08-13T07:21:16.945845542Z" level=info msg="Start event monitor" Aug 13 07:21:16.945870 containerd[1920]: time="2025-08-13T07:21:16.945851195Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 07:21:16.945870 containerd[1920]: time="2025-08-13T07:21:16.945852534Z" level=info msg="Start snapshots syncer" Aug 13 07:21:16.945903 containerd[1920]: time="2025-08-13T07:21:16.945869985Z" level=info msg="Start cni network conf syncer for default" Aug 13 07:21:16.945903 containerd[1920]: time="2025-08-13T07:21:16.945874506Z" level=info msg="Start streaming server" Aug 13 07:21:16.945903 containerd[1920]: time="2025-08-13T07:21:16.945879083Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 07:21:16.945949 containerd[1920]: time="2025-08-13T07:21:16.945913496Z" level=info msg="containerd successfully booted in 0.029705s" Aug 13 07:21:16.949302 systemd[1]: Reached target getty.target - Login Prompts. Aug 13 07:21:16.957388 systemd[1]: Started containerd.service - containerd container runtime. Aug 13 07:21:16.996101 kernel: EXT4-fs (sdb9): resized filesystem to 116605649 Aug 13 07:21:17.016848 extend-filesystems[1893]: Filesystem at /dev/sdb9 is mounted on /; on-line resizing required Aug 13 07:21:17.016848 extend-filesystems[1893]: old_desc_blocks = 1, new_desc_blocks = 56 Aug 13 07:21:17.016848 extend-filesystems[1893]: The filesystem on /dev/sdb9 is now 116605649 (4k) blocks long. Aug 13 07:21:17.017663 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 07:21:17.057216 tar[1918]: linux-amd64/LICENSE Aug 13 07:21:17.057216 tar[1918]: linux-amd64/README.md Aug 13 07:21:17.057256 extend-filesystems[1884]: Resized filesystem in /dev/sdb9 Aug 13 07:21:17.017827 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Aug 13 07:21:17.068180 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Aug 13 07:21:17.610319 systemd-networkd[1561]: bond0: Gained IPv6LL Aug 13 07:21:17.611976 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Aug 13 07:21:17.624274 systemd[1]: Reached target network-online.target - Network is Online. Aug 13 07:21:17.658486 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:21:17.669196 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Aug 13 07:21:17.694249 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Aug 13 07:21:18.397912 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:18.410689 (kubelet)[2034]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:21:18.828018 kubelet[2034]: E0813 07:21:18.827934 2034 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:21:18.829168 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:21:18.829254 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:21:19.297410 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Aug 13 07:21:19.314321 systemd[1]: Started sshd@0-145.40.90.243:22-147.75.109.163:59524.service - OpenSSH per-connection server daemon (147.75.109.163:59524). Aug 13 07:21:19.350957 systemd-timesyncd[1872]: Contacted time server 50.218.103.254:123 (0.flatcar.pool.ntp.org). Aug 13 07:21:19.351004 systemd-timesyncd[1872]: Initial clock synchronization to Wed 2025-08-13 07:21:19.585634 UTC. Aug 13 07:21:19.360669 sshd[2053]: Accepted publickey for core from 147.75.109.163 port 59524 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:19.361971 sshd[2053]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:19.367478 systemd-logind[1908]: New session 1 of user core. Aug 13 07:21:19.368246 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Aug 13 07:21:19.391257 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Aug 13 07:21:19.401709 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Aug 13 07:21:19.431386 systemd[1]: Starting user@500.service - User Manager for UID 500... Aug 13 07:21:19.447710 (systemd)[2059]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 07:21:19.520910 systemd[2059]: Queued start job for default target default.target. Aug 13 07:21:19.521088 systemd[2059]: Created slice app.slice - User Application Slice. Aug 13 07:21:19.521100 systemd[2059]: Reached target paths.target - Paths. Aug 13 07:21:19.521109 systemd[2059]: Reached target timers.target - Timers. Aug 13 07:21:19.535156 systemd[2059]: Starting dbus.socket - D-Bus User Message Bus Socket... Aug 13 07:21:19.538490 systemd[2059]: Listening on dbus.socket - D-Bus User Message Bus Socket. Aug 13 07:21:19.538519 systemd[2059]: Reached target sockets.target - Sockets. Aug 13 07:21:19.538529 systemd[2059]: Reached target basic.target - Basic System. Aug 13 07:21:19.538550 systemd[2059]: Reached target default.target - Main User Target. Aug 13 07:21:19.538565 systemd[2059]: Startup finished in 86ms. Aug 13 07:21:19.538733 systemd[1]: Started user@500.service - User Manager for UID 500. Aug 13 07:21:19.549118 systemd[1]: Started session-1.scope - Session 1 of User core. Aug 13 07:21:19.622127 kernel: mlx5_core 0000:01:00.0: lag map: port 1:1 port 2:2 Aug 13 07:21:19.622276 kernel: mlx5_core 0000:01:00.0: shared_fdb:0 mode:queue_affinity Aug 13 07:21:19.650297 systemd[1]: Started sshd@1-145.40.90.243:22-147.75.109.163:55112.service - OpenSSH per-connection server daemon (147.75.109.163:55112). Aug 13 07:21:19.679262 sshd[2071]: Accepted publickey for core from 147.75.109.163 port 55112 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:19.685415 sshd[2071]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:19.697492 systemd-logind[1908]: New session 2 of user core. Aug 13 07:21:19.715900 systemd[1]: Started session-2.scope - Session 2 of User core. Aug 13 07:21:19.794652 sshd[2071]: pam_unix(sshd:session): session closed for user core Aug 13 07:21:19.815177 systemd[1]: Started sshd@2-145.40.90.243:22-147.75.109.163:55124.service - OpenSSH per-connection server daemon (147.75.109.163:55124). Aug 13 07:21:19.828337 systemd[1]: sshd@1-145.40.90.243:22-147.75.109.163:55112.service: Deactivated successfully. Aug 13 07:21:19.829216 systemd[1]: session-2.scope: Deactivated successfully. Aug 13 07:21:19.829672 systemd-logind[1908]: Session 2 logged out. Waiting for processes to exit. Aug 13 07:21:19.830325 systemd-logind[1908]: Removed session 2. Aug 13 07:21:19.853672 sshd[2081]: Accepted publickey for core from 147.75.109.163 port 55124 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:19.854338 sshd[2081]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:19.856618 systemd-logind[1908]: New session 3 of user core. Aug 13 07:21:19.869891 systemd[1]: Started session-3.scope - Session 3 of User core. Aug 13 07:21:19.933509 sshd[2081]: pam_unix(sshd:session): session closed for user core Aug 13 07:21:19.934670 systemd[1]: sshd@2-145.40.90.243:22-147.75.109.163:55124.service: Deactivated successfully. Aug 13 07:21:19.935851 systemd-logind[1908]: Session 3 logged out. Waiting for processes to exit. Aug 13 07:21:19.935956 systemd[1]: session-3.scope: Deactivated successfully. Aug 13 07:21:19.936521 systemd-logind[1908]: Removed session 3. Aug 13 07:21:20.279228 coreos-metadata[1984]: Aug 13 07:21:20.279 INFO Fetch successful Aug 13 07:21:20.311872 coreos-metadata[1878]: Aug 13 07:21:20.311 INFO Fetch successful Aug 13 07:21:20.314065 unknown[1984]: wrote ssh authorized keys file for user: core Aug 13 07:21:20.333741 update-ssh-keys[2093]: Updated "/home/core/.ssh/authorized_keys" Aug 13 07:21:20.334585 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Aug 13 07:21:20.347254 systemd[1]: Finished sshkeys.service. Aug 13 07:21:20.359946 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Aug 13 07:21:20.381361 systemd[1]: Starting packet-phone-home.service - Report Success to Packet... Aug 13 07:21:20.798992 systemd[1]: Finished packet-phone-home.service - Report Success to Packet. Aug 13 07:21:20.811153 systemd[1]: Reached target multi-user.target - Multi-User System. Aug 13 07:21:20.821658 systemd[1]: Startup finished in 27.519s (kernel) + 10.200s (userspace) = 37.719s. Aug 13 07:21:20.859681 login[2000]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Aug 13 07:21:20.862832 systemd-logind[1908]: New session 4 of user core. Aug 13 07:21:20.863373 systemd[1]: Started session-4.scope - Session 4 of User core. Aug 13 07:21:20.864906 login[1997]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Aug 13 07:21:20.867345 systemd-logind[1908]: New session 5 of user core. Aug 13 07:21:20.867853 systemd[1]: Started session-5.scope - Session 5 of User core. Aug 13 07:21:29.091321 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 07:21:29.107399 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:21:29.349103 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:29.354089 (kubelet)[2151]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:21:29.386127 kubelet[2151]: E0813 07:21:29.386053 2151 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:21:29.388235 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:21:29.388329 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:21:30.123351 systemd[1]: Started sshd@3-145.40.90.243:22-147.75.109.163:36174.service - OpenSSH per-connection server daemon (147.75.109.163:36174). Aug 13 07:21:30.153609 sshd[2170]: Accepted publickey for core from 147.75.109.163 port 36174 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:30.154725 sshd[2170]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:30.158705 systemd-logind[1908]: New session 6 of user core. Aug 13 07:21:30.167496 systemd[1]: Started session-6.scope - Session 6 of User core. Aug 13 07:21:30.223700 sshd[2170]: pam_unix(sshd:session): session closed for user core Aug 13 07:21:30.233376 systemd[1]: Started sshd@4-145.40.90.243:22-147.75.109.163:36180.service - OpenSSH per-connection server daemon (147.75.109.163:36180). Aug 13 07:21:30.233723 systemd[1]: sshd@3-145.40.90.243:22-147.75.109.163:36174.service: Deactivated successfully. Aug 13 07:21:30.234491 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 07:21:30.234876 systemd-logind[1908]: Session 6 logged out. Waiting for processes to exit. Aug 13 07:21:30.235783 systemd-logind[1908]: Removed session 6. Aug 13 07:21:30.266235 sshd[2176]: Accepted publickey for core from 147.75.109.163 port 36180 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:30.269682 sshd[2176]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:30.281403 systemd-logind[1908]: New session 7 of user core. Aug 13 07:21:30.300028 systemd[1]: Started session-7.scope - Session 7 of User core. Aug 13 07:21:30.361388 sshd[2176]: pam_unix(sshd:session): session closed for user core Aug 13 07:21:30.376768 systemd[1]: Started sshd@5-145.40.90.243:22-147.75.109.163:36182.service - OpenSSH per-connection server daemon (147.75.109.163:36182). Aug 13 07:21:30.378481 systemd[1]: sshd@4-145.40.90.243:22-147.75.109.163:36180.service: Deactivated successfully. Aug 13 07:21:30.382418 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 07:21:30.384308 systemd-logind[1908]: Session 7 logged out. Waiting for processes to exit. Aug 13 07:21:30.387898 systemd-logind[1908]: Removed session 7. Aug 13 07:21:30.436506 sshd[2186]: Accepted publickey for core from 147.75.109.163 port 36182 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:30.438299 sshd[2186]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:30.444654 systemd-logind[1908]: New session 8 of user core. Aug 13 07:21:30.462940 systemd[1]: Started session-8.scope - Session 8 of User core. Aug 13 07:21:30.524259 sshd[2186]: pam_unix(sshd:session): session closed for user core Aug 13 07:21:30.534402 systemd[1]: Started sshd@6-145.40.90.243:22-147.75.109.163:36192.service - OpenSSH per-connection server daemon (147.75.109.163:36192). Aug 13 07:21:30.534741 systemd[1]: sshd@5-145.40.90.243:22-147.75.109.163:36182.service: Deactivated successfully. Aug 13 07:21:30.535532 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 07:21:30.536315 systemd-logind[1908]: Session 8 logged out. Waiting for processes to exit. Aug 13 07:21:30.537020 systemd-logind[1908]: Removed session 8. Aug 13 07:21:30.565653 sshd[2193]: Accepted publickey for core from 147.75.109.163 port 36192 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:30.568996 sshd[2193]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:30.580791 systemd-logind[1908]: New session 9 of user core. Aug 13 07:21:30.591818 systemd[1]: Started session-9.scope - Session 9 of User core. Aug 13 07:21:30.677954 sudo[2200]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Aug 13 07:21:30.678110 sudo[2200]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:21:30.695928 sudo[2200]: pam_unix(sudo:session): session closed for user root Aug 13 07:21:30.697001 sshd[2193]: pam_unix(sshd:session): session closed for user core Aug 13 07:21:30.718421 systemd[1]: Started sshd@7-145.40.90.243:22-147.75.109.163:36198.service - OpenSSH per-connection server daemon (147.75.109.163:36198). Aug 13 07:21:30.718933 systemd[1]: sshd@6-145.40.90.243:22-147.75.109.163:36192.service: Deactivated successfully. Aug 13 07:21:30.721095 systemd-logind[1908]: Session 9 logged out. Waiting for processes to exit. Aug 13 07:21:30.721227 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 07:21:30.722448 systemd-logind[1908]: Removed session 9. Aug 13 07:21:30.750410 sshd[2202]: Accepted publickey for core from 147.75.109.163 port 36198 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:30.753836 sshd[2202]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:30.765186 systemd-logind[1908]: New session 10 of user core. Aug 13 07:21:30.781798 systemd[1]: Started session-10.scope - Session 10 of User core. Aug 13 07:21:30.851664 sudo[2210]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Aug 13 07:21:30.852554 sudo[2210]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:21:30.861353 sudo[2210]: pam_unix(sudo:session): session closed for user root Aug 13 07:21:30.875417 sudo[2209]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Aug 13 07:21:30.876258 sudo[2209]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:21:30.919877 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Aug 13 07:21:30.923575 auditctl[2213]: No rules Aug 13 07:21:30.924502 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 07:21:30.925135 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Aug 13 07:21:30.931515 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 07:21:31.005764 augenrules[2232]: No rules Aug 13 07:21:31.007569 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 07:21:31.009962 sudo[2209]: pam_unix(sudo:session): session closed for user root Aug 13 07:21:31.013454 sshd[2202]: pam_unix(sshd:session): session closed for user core Aug 13 07:21:31.021003 systemd[1]: sshd@7-145.40.90.243:22-147.75.109.163:36198.service: Deactivated successfully. Aug 13 07:21:31.027026 systemd-logind[1908]: Session 10 logged out. Waiting for processes to exit. Aug 13 07:21:31.047912 systemd[1]: Started sshd@8-145.40.90.243:22-147.75.109.163:36212.service - OpenSSH per-connection server daemon (147.75.109.163:36212). Aug 13 07:21:31.049427 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 07:21:31.052712 systemd-logind[1908]: Removed session 10. Aug 13 07:21:31.104629 sshd[2241]: Accepted publickey for core from 147.75.109.163 port 36212 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:21:31.106041 sshd[2241]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:21:31.111003 systemd-logind[1908]: New session 11 of user core. Aug 13 07:21:31.121532 systemd[1]: Started session-11.scope - Session 11 of User core. Aug 13 07:21:31.186668 sudo[2245]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 07:21:31.187510 sudo[2245]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:21:31.568422 systemd[1]: Starting docker.service - Docker Application Container Engine... Aug 13 07:21:31.568549 (dockerd)[2270]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Aug 13 07:21:31.808886 dockerd[2270]: time="2025-08-13T07:21:31.808828834Z" level=info msg="Starting up" Aug 13 07:21:32.062211 dockerd[2270]: time="2025-08-13T07:21:32.062134110Z" level=info msg="Loading containers: start." Aug 13 07:21:32.150118 kernel: Initializing XFRM netlink socket Aug 13 07:21:32.202445 systemd-networkd[1561]: docker0: Link UP Aug 13 07:21:32.215013 dockerd[2270]: time="2025-08-13T07:21:32.214992764Z" level=info msg="Loading containers: done." Aug 13 07:21:32.224715 dockerd[2270]: time="2025-08-13T07:21:32.224657091Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 07:21:32.224715 dockerd[2270]: time="2025-08-13T07:21:32.224708394Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Aug 13 07:21:32.224803 dockerd[2270]: time="2025-08-13T07:21:32.224759889Z" level=info msg="Daemon has completed initialization" Aug 13 07:21:32.256789 dockerd[2270]: time="2025-08-13T07:21:32.256738554Z" level=info msg="API listen on /run/docker.sock" Aug 13 07:21:32.256900 systemd[1]: Started docker.service - Docker Application Container Engine. Aug 13 07:21:33.179059 containerd[1920]: time="2025-08-13T07:21:33.179010066Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\"" Aug 13 07:21:34.011816 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2958163605.mount: Deactivated successfully. Aug 13 07:21:34.715832 containerd[1920]: time="2025-08-13T07:21:34.715779220Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:34.716083 containerd[1920]: time="2025-08-13T07:21:34.715947046Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.11: active requests=0, bytes read=28077759" Aug 13 07:21:34.716405 containerd[1920]: time="2025-08-13T07:21:34.716365679Z" level=info msg="ImageCreate event name:\"sha256:ea7fa3cfabed1b85e7de8e0a02356b6dcb7708442d6e4600d68abaebe1e9b1fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:34.718512 containerd[1920]: time="2025-08-13T07:21:34.718479198Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:a3d1c4440817725a1b503a7ccce94f3dce2b208ebf257b405dc2d97817df3dde\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:34.719359 containerd[1920]: time="2025-08-13T07:21:34.719314154Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.11\" with image id \"sha256:ea7fa3cfabed1b85e7de8e0a02356b6dcb7708442d6e4600d68abaebe1e9b1fc\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:a3d1c4440817725a1b503a7ccce94f3dce2b208ebf257b405dc2d97817df3dde\", size \"28074559\" in 1.540280734s" Aug 13 07:21:34.719359 containerd[1920]: time="2025-08-13T07:21:34.719333027Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\" returns image reference \"sha256:ea7fa3cfabed1b85e7de8e0a02356b6dcb7708442d6e4600d68abaebe1e9b1fc\"" Aug 13 07:21:34.719658 containerd[1920]: time="2025-08-13T07:21:34.719646367Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\"" Aug 13 07:21:35.799461 containerd[1920]: time="2025-08-13T07:21:35.799435813Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:35.799704 containerd[1920]: time="2025-08-13T07:21:35.799618749Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.11: active requests=0, bytes read=24713245" Aug 13 07:21:35.800053 containerd[1920]: time="2025-08-13T07:21:35.800040785Z" level=info msg="ImageCreate event name:\"sha256:c057eceea4b436b01f9ce394734cfb06f13b2a3688c3983270e99743370b6051\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:35.801586 containerd[1920]: time="2025-08-13T07:21:35.801572021Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:0f19de157f3d251f5ddeb6e9d026895bc55cb02592874b326fa345c57e5e2848\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:35.802267 containerd[1920]: time="2025-08-13T07:21:35.802230357Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.11\" with image id \"sha256:c057eceea4b436b01f9ce394734cfb06f13b2a3688c3983270e99743370b6051\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:0f19de157f3d251f5ddeb6e9d026895bc55cb02592874b326fa345c57e5e2848\", size \"26315079\" in 1.082566206s" Aug 13 07:21:35.802267 containerd[1920]: time="2025-08-13T07:21:35.802247422Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\" returns image reference \"sha256:c057eceea4b436b01f9ce394734cfb06f13b2a3688c3983270e99743370b6051\"" Aug 13 07:21:35.802549 containerd[1920]: time="2025-08-13T07:21:35.802513958Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\"" Aug 13 07:21:36.680042 containerd[1920]: time="2025-08-13T07:21:36.679986264Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:36.680277 containerd[1920]: time="2025-08-13T07:21:36.680228308Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.11: active requests=0, bytes read=18783700" Aug 13 07:21:36.680599 containerd[1920]: time="2025-08-13T07:21:36.680559570Z" level=info msg="ImageCreate event name:\"sha256:64e6a0b453108c87da0bb61473b35fd54078119a09edc56a4c8cb31602437c58\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:36.682122 containerd[1920]: time="2025-08-13T07:21:36.682065328Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:1a9b59b3bfa6c1f1911f6f865a795620c461d079e413061bb71981cadd67f39d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:36.682676 containerd[1920]: time="2025-08-13T07:21:36.682637545Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.11\" with image id \"sha256:64e6a0b453108c87da0bb61473b35fd54078119a09edc56a4c8cb31602437c58\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:1a9b59b3bfa6c1f1911f6f865a795620c461d079e413061bb71981cadd67f39d\", size \"20385552\" in 880.107452ms" Aug 13 07:21:36.682676 containerd[1920]: time="2025-08-13T07:21:36.682653518Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\" returns image reference \"sha256:64e6a0b453108c87da0bb61473b35fd54078119a09edc56a4c8cb31602437c58\"" Aug 13 07:21:36.682926 containerd[1920]: time="2025-08-13T07:21:36.682908942Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\"" Aug 13 07:21:37.520581 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount58792575.mount: Deactivated successfully. Aug 13 07:21:37.708433 containerd[1920]: time="2025-08-13T07:21:37.708405426Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:37.708736 containerd[1920]: time="2025-08-13T07:21:37.708618892Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.11: active requests=0, bytes read=30383612" Aug 13 07:21:37.708915 containerd[1920]: time="2025-08-13T07:21:37.708902746Z" level=info msg="ImageCreate event name:\"sha256:0cec28fd5c3c446ec52e2886ddea38bf7f7e17755aa5d0095d50d3df5914a8fd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:37.709926 containerd[1920]: time="2025-08-13T07:21:37.709912046Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:a31da847792c5e7e92e91b78da1ad21d693e4b2b48d0e9f4610c8764dc2a5d79\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:37.710404 containerd[1920]: time="2025-08-13T07:21:37.710358411Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.11\" with image id \"sha256:0cec28fd5c3c446ec52e2886ddea38bf7f7e17755aa5d0095d50d3df5914a8fd\", repo tag \"registry.k8s.io/kube-proxy:v1.31.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:a31da847792c5e7e92e91b78da1ad21d693e4b2b48d0e9f4610c8764dc2a5d79\", size \"30382631\" in 1.027433595s" Aug 13 07:21:37.710404 containerd[1920]: time="2025-08-13T07:21:37.710378494Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\" returns image reference \"sha256:0cec28fd5c3c446ec52e2886ddea38bf7f7e17755aa5d0095d50d3df5914a8fd\"" Aug 13 07:21:37.710683 containerd[1920]: time="2025-08-13T07:21:37.710671943Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Aug 13 07:21:38.269572 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2795289760.mount: Deactivated successfully. Aug 13 07:21:38.821425 containerd[1920]: time="2025-08-13T07:21:38.821400079Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:38.821728 containerd[1920]: time="2025-08-13T07:21:38.821571731Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Aug 13 07:21:38.822129 containerd[1920]: time="2025-08-13T07:21:38.822115209Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:38.824140 containerd[1920]: time="2025-08-13T07:21:38.824094077Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:38.824698 containerd[1920]: time="2025-08-13T07:21:38.824682157Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.113994545s" Aug 13 07:21:38.824742 containerd[1920]: time="2025-08-13T07:21:38.824699542Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Aug 13 07:21:38.824950 containerd[1920]: time="2025-08-13T07:21:38.824935840Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 07:21:39.262917 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount530315237.mount: Deactivated successfully. Aug 13 07:21:39.264176 containerd[1920]: time="2025-08-13T07:21:39.264099212Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:39.264378 containerd[1920]: time="2025-08-13T07:21:39.264331880Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Aug 13 07:21:39.264631 containerd[1920]: time="2025-08-13T07:21:39.264603331Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:39.265860 containerd[1920]: time="2025-08-13T07:21:39.265824755Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:39.266350 containerd[1920]: time="2025-08-13T07:21:39.266307914Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 441.352731ms" Aug 13 07:21:39.266350 containerd[1920]: time="2025-08-13T07:21:39.266322490Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Aug 13 07:21:39.266866 containerd[1920]: time="2025-08-13T07:21:39.266797305Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Aug 13 07:21:39.531013 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 07:21:39.549461 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:21:39.798245 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:39.800724 (kubelet)[2577]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:21:39.821652 kubelet[2577]: E0813 07:21:39.821614 2577 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:21:39.822810 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:21:39.822902 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:21:39.831606 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount48159507.mount: Deactivated successfully. Aug 13 07:21:40.873321 containerd[1920]: time="2025-08-13T07:21:40.873267263Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:40.873549 containerd[1920]: time="2025-08-13T07:21:40.873485631Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56780013" Aug 13 07:21:40.873911 containerd[1920]: time="2025-08-13T07:21:40.873891659Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:40.875879 containerd[1920]: time="2025-08-13T07:21:40.875864046Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:40.876474 containerd[1920]: time="2025-08-13T07:21:40.876459949Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 1.609614323s" Aug 13 07:21:40.876514 containerd[1920]: time="2025-08-13T07:21:40.876475811Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Aug 13 07:21:42.755962 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:42.771401 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:21:42.783160 systemd[1]: Reloading requested from client PID 2705 ('systemctl') (unit session-11.scope)... Aug 13 07:21:42.783168 systemd[1]: Reloading... Aug 13 07:21:42.818136 zram_generator::config[2744]: No configuration found. Aug 13 07:21:42.898497 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:21:42.959348 systemd[1]: Reloading finished in 175 ms. Aug 13 07:21:43.004249 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Aug 13 07:21:43.004567 systemd[1]: kubelet.service: Failed with result 'signal'. Aug 13 07:21:43.005294 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:43.019990 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:21:43.316837 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:43.319259 (kubelet)[2818]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:21:43.340632 kubelet[2818]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:21:43.340632 kubelet[2818]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 07:21:43.340632 kubelet[2818]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:21:43.340632 kubelet[2818]: I0813 07:21:43.340627 2818 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:21:43.487834 kubelet[2818]: I0813 07:21:43.487818 2818 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 07:21:43.487834 kubelet[2818]: I0813 07:21:43.487829 2818 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:21:43.487959 kubelet[2818]: I0813 07:21:43.487953 2818 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 07:21:43.506738 kubelet[2818]: E0813 07:21:43.506700 2818 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://145.40.90.243:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:43.510713 kubelet[2818]: I0813 07:21:43.510683 2818 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:21:43.516923 kubelet[2818]: E0813 07:21:43.516908 2818 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:21:43.516923 kubelet[2818]: I0813 07:21:43.516924 2818 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:21:43.527729 kubelet[2818]: I0813 07:21:43.527705 2818 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:21:43.528423 kubelet[2818]: I0813 07:21:43.528387 2818 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 07:21:43.528467 kubelet[2818]: I0813 07:21:43.528449 2818 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:21:43.528585 kubelet[2818]: I0813 07:21:43.528465 2818 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.5-a-3ca3f1276f","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Aug 13 07:21:43.528585 kubelet[2818]: I0813 07:21:43.528559 2818 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:21:43.528585 kubelet[2818]: I0813 07:21:43.528565 2818 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 07:21:43.528679 kubelet[2818]: I0813 07:21:43.528617 2818 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:21:43.530842 kubelet[2818]: I0813 07:21:43.530812 2818 kubelet.go:408] "Attempting to sync node with API server" Aug 13 07:21:43.530842 kubelet[2818]: I0813 07:21:43.530840 2818 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:21:43.530935 kubelet[2818]: I0813 07:21:43.530857 2818 kubelet.go:314] "Adding apiserver pod source" Aug 13 07:21:43.530935 kubelet[2818]: I0813 07:21:43.530866 2818 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:21:43.532846 kubelet[2818]: I0813 07:21:43.532833 2818 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 07:21:43.533255 kubelet[2818]: I0813 07:21:43.533208 2818 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 07:21:43.533581 kubelet[2818]: W0813 07:21:43.533538 2818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://145.40.90.243:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.5-a-3ca3f1276f&limit=500&resourceVersion=0": dial tcp 145.40.90.243:6443: connect: connection refused Aug 13 07:21:43.533629 kubelet[2818]: E0813 07:21:43.533592 2818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://145.40.90.243:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.5-a-3ca3f1276f&limit=500&resourceVersion=0\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:43.534033 kubelet[2818]: W0813 07:21:43.534015 2818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://145.40.90.243:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 145.40.90.243:6443: connect: connection refused Aug 13 07:21:43.534058 kubelet[2818]: E0813 07:21:43.534039 2818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://145.40.90.243:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:43.534122 kubelet[2818]: W0813 07:21:43.534061 2818 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 07:21:43.535827 kubelet[2818]: I0813 07:21:43.535782 2818 server.go:1274] "Started kubelet" Aug 13 07:21:43.535928 kubelet[2818]: I0813 07:21:43.535846 2818 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:21:43.535973 kubelet[2818]: I0813 07:21:43.535937 2818 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:21:43.536212 kubelet[2818]: I0813 07:21:43.536179 2818 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:21:43.536987 kubelet[2818]: I0813 07:21:43.536980 2818 server.go:449] "Adding debug handlers to kubelet server" Aug 13 07:21:43.537013 kubelet[2818]: I0813 07:21:43.537006 2818 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:21:43.537038 kubelet[2818]: I0813 07:21:43.537014 2818 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:21:43.537124 kubelet[2818]: E0813 07:21:43.537100 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:43.537458 kubelet[2818]: I0813 07:21:43.537438 2818 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 07:21:43.537912 kubelet[2818]: I0813 07:21:43.537472 2818 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 07:21:43.538055 kubelet[2818]: I0813 07:21:43.538044 2818 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:21:43.538519 kubelet[2818]: E0813 07:21:43.538490 2818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://145.40.90.243:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-a-3ca3f1276f?timeout=10s\": dial tcp 145.40.90.243:6443: connect: connection refused" interval="200ms" Aug 13 07:21:43.538569 kubelet[2818]: W0813 07:21:43.538501 2818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://145.40.90.243:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 145.40.90.243:6443: connect: connection refused Aug 13 07:21:43.538569 kubelet[2818]: E0813 07:21:43.538545 2818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://145.40.90.243:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:43.538569 kubelet[2818]: I0813 07:21:43.538565 2818 factory.go:221] Registration of the systemd container factory successfully Aug 13 07:21:43.538657 kubelet[2818]: I0813 07:21:43.538645 2818 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:21:43.541210 kubelet[2818]: E0813 07:21:43.541200 2818 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:21:43.541279 kubelet[2818]: I0813 07:21:43.541272 2818 factory.go:221] Registration of the containerd container factory successfully Aug 13 07:21:43.542239 kubelet[2818]: E0813 07:21:43.541154 2818 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://145.40.90.243:6443/api/v1/namespaces/default/events\": dial tcp 145.40.90.243:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.5-a-3ca3f1276f.185b4298bb4caee5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.5-a-3ca3f1276f,UID:ci-4081.3.5-a-3ca3f1276f,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.5-a-3ca3f1276f,},FirstTimestamp:2025-08-13 07:21:43.535767269 +0000 UTC m=+0.214613936,LastTimestamp:2025-08-13 07:21:43.535767269 +0000 UTC m=+0.214613936,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.5-a-3ca3f1276f,}" Aug 13 07:21:43.547567 kubelet[2818]: I0813 07:21:43.547549 2818 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 07:21:43.548058 kubelet[2818]: I0813 07:21:43.548048 2818 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 07:21:43.548113 kubelet[2818]: I0813 07:21:43.548061 2818 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 07:21:43.548113 kubelet[2818]: I0813 07:21:43.548079 2818 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 07:21:43.548155 kubelet[2818]: E0813 07:21:43.548110 2818 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:21:43.548294 kubelet[2818]: W0813 07:21:43.548278 2818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://145.40.90.243:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 145.40.90.243:6443: connect: connection refused Aug 13 07:21:43.548317 kubelet[2818]: E0813 07:21:43.548301 2818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://145.40.90.243:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:43.577686 kubelet[2818]: I0813 07:21:43.577640 2818 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 07:21:43.577686 kubelet[2818]: I0813 07:21:43.577651 2818 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 07:21:43.577686 kubelet[2818]: I0813 07:21:43.577664 2818 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:21:43.578622 kubelet[2818]: I0813 07:21:43.578613 2818 policy_none.go:49] "None policy: Start" Aug 13 07:21:43.578872 kubelet[2818]: I0813 07:21:43.578862 2818 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 07:21:43.578915 kubelet[2818]: I0813 07:21:43.578876 2818 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:21:43.581357 kubelet[2818]: I0813 07:21:43.581341 2818 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 07:21:43.581592 kubelet[2818]: I0813 07:21:43.581546 2818 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:21:43.581631 kubelet[2818]: I0813 07:21:43.581598 2818 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:21:43.581720 kubelet[2818]: I0813 07:21:43.581712 2818 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:21:43.582322 kubelet[2818]: E0813 07:21:43.582315 2818 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:43.684085 kubelet[2818]: I0813 07:21:43.684034 2818 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.684387 kubelet[2818]: E0813 07:21:43.684338 2818 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://145.40.90.243:6443/api/v1/nodes\": dial tcp 145.40.90.243:6443: connect: connection refused" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.739824 kubelet[2818]: E0813 07:21:43.739696 2818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://145.40.90.243:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-a-3ca3f1276f?timeout=10s\": dial tcp 145.40.90.243:6443: connect: connection refused" interval="400ms" Aug 13 07:21:43.840501 kubelet[2818]: I0813 07:21:43.840230 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.840501 kubelet[2818]: I0813 07:21:43.840358 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b2c458f081aa2526b69831ffeb469bd5-kubeconfig\") pod \"kube-scheduler-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b2c458f081aa2526b69831ffeb469bd5\") " pod="kube-system/kube-scheduler-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.840501 kubelet[2818]: I0813 07:21:43.840450 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b57a63056e7190c205656dc6817567bc-ca-certs\") pod \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b57a63056e7190c205656dc6817567bc\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.840927 kubelet[2818]: I0813 07:21:43.840509 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b57a63056e7190c205656dc6817567bc-k8s-certs\") pod \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b57a63056e7190c205656dc6817567bc\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.840927 kubelet[2818]: I0813 07:21:43.840574 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-ca-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.840927 kubelet[2818]: I0813 07:21:43.840643 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.840927 kubelet[2818]: I0813 07:21:43.840707 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b57a63056e7190c205656dc6817567bc-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b57a63056e7190c205656dc6817567bc\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.840927 kubelet[2818]: I0813 07:21:43.840759 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.841409 kubelet[2818]: I0813 07:21:43.840821 2818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.889579 kubelet[2818]: I0813 07:21:43.889472 2818 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.890328 kubelet[2818]: E0813 07:21:43.890224 2818 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://145.40.90.243:6443/api/v1/nodes\": dial tcp 145.40.90.243:6443: connect: connection refused" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:43.953396 containerd[1920]: time="2025-08-13T07:21:43.953272016Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.5-a-3ca3f1276f,Uid:b2c458f081aa2526b69831ffeb469bd5,Namespace:kube-system,Attempt:0,}" Aug 13 07:21:43.953396 containerd[1920]: time="2025-08-13T07:21:43.953338654Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.5-a-3ca3f1276f,Uid:b57a63056e7190c205656dc6817567bc,Namespace:kube-system,Attempt:0,}" Aug 13 07:21:43.954323 containerd[1920]: time="2025-08-13T07:21:43.953925862Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.5-a-3ca3f1276f,Uid:18e75a8702129840c848c326f4a5a54c,Namespace:kube-system,Attempt:0,}" Aug 13 07:21:44.141047 kubelet[2818]: E0813 07:21:44.140788 2818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://145.40.90.243:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-a-3ca3f1276f?timeout=10s\": dial tcp 145.40.90.243:6443: connect: connection refused" interval="800ms" Aug 13 07:21:44.164556 kubelet[2818]: E0813 07:21:44.164296 2818 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://145.40.90.243:6443/api/v1/namespaces/default/events\": dial tcp 145.40.90.243:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.5-a-3ca3f1276f.185b4298bb4caee5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.5-a-3ca3f1276f,UID:ci-4081.3.5-a-3ca3f1276f,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.5-a-3ca3f1276f,},FirstTimestamp:2025-08-13 07:21:43.535767269 +0000 UTC m=+0.214613936,LastTimestamp:2025-08-13 07:21:43.535767269 +0000 UTC m=+0.214613936,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.5-a-3ca3f1276f,}" Aug 13 07:21:44.292287 kubelet[2818]: I0813 07:21:44.292265 2818 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:44.292573 kubelet[2818]: E0813 07:21:44.292524 2818 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://145.40.90.243:6443/api/v1/nodes\": dial tcp 145.40.90.243:6443: connect: connection refused" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:44.413951 kubelet[2818]: W0813 07:21:44.413829 2818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://145.40.90.243:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 145.40.90.243:6443: connect: connection refused Aug 13 07:21:44.413951 kubelet[2818]: E0813 07:21:44.413885 2818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://145.40.90.243:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:44.446923 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3477029806.mount: Deactivated successfully. Aug 13 07:21:44.448457 containerd[1920]: time="2025-08-13T07:21:44.448435995Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:21:44.448686 containerd[1920]: time="2025-08-13T07:21:44.448666403Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Aug 13 07:21:44.449050 containerd[1920]: time="2025-08-13T07:21:44.449039785Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:21:44.449520 containerd[1920]: time="2025-08-13T07:21:44.449481440Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:21:44.449520 containerd[1920]: time="2025-08-13T07:21:44.449497308Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:21:44.449989 containerd[1920]: time="2025-08-13T07:21:44.449948761Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:21:44.450787 containerd[1920]: time="2025-08-13T07:21:44.450738535Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:21:44.451338 containerd[1920]: time="2025-08-13T07:21:44.451297978Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:21:44.452406 containerd[1920]: time="2025-08-13T07:21:44.452361495Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 498.261937ms" Aug 13 07:21:44.453069 containerd[1920]: time="2025-08-13T07:21:44.453026150Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 499.545551ms" Aug 13 07:21:44.454492 containerd[1920]: time="2025-08-13T07:21:44.454453266Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 501.009577ms" Aug 13 07:21:44.463250 kubelet[2818]: W0813 07:21:44.463190 2818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://145.40.90.243:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.5-a-3ca3f1276f&limit=500&resourceVersion=0": dial tcp 145.40.90.243:6443: connect: connection refused Aug 13 07:21:44.463291 kubelet[2818]: E0813 07:21:44.463260 2818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://145.40.90.243:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.5-a-3ca3f1276f&limit=500&resourceVersion=0\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:44.533724 containerd[1920]: time="2025-08-13T07:21:44.533679393Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:21:44.533724 containerd[1920]: time="2025-08-13T07:21:44.533708705Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:21:44.533724 containerd[1920]: time="2025-08-13T07:21:44.533715866Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:44.533853 containerd[1920]: time="2025-08-13T07:21:44.533761846Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:44.533853 containerd[1920]: time="2025-08-13T07:21:44.533579158Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:21:44.533853 containerd[1920]: time="2025-08-13T07:21:44.533791634Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:21:44.533853 containerd[1920]: time="2025-08-13T07:21:44.533800568Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:44.533966 containerd[1920]: time="2025-08-13T07:21:44.533845591Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:44.534000 containerd[1920]: time="2025-08-13T07:21:44.533947481Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:21:44.534179 containerd[1920]: time="2025-08-13T07:21:44.534164200Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:21:44.534217 containerd[1920]: time="2025-08-13T07:21:44.534175758Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:44.534246 containerd[1920]: time="2025-08-13T07:21:44.534233177Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:44.542778 kubelet[2818]: W0813 07:21:44.542712 2818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://145.40.90.243:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 145.40.90.243:6443: connect: connection refused Aug 13 07:21:44.542778 kubelet[2818]: E0813 07:21:44.542757 2818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://145.40.90.243:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 145.40.90.243:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:21:44.573050 containerd[1920]: time="2025-08-13T07:21:44.573023853Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.5-a-3ca3f1276f,Uid:18e75a8702129840c848c326f4a5a54c,Namespace:kube-system,Attempt:0,} returns sandbox id \"05512df33da93524185c8a95df9f2d46a944cb07ea60992b1e60a7e539625732\"" Aug 13 07:21:44.573050 containerd[1920]: time="2025-08-13T07:21:44.573052382Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.5-a-3ca3f1276f,Uid:b2c458f081aa2526b69831ffeb469bd5,Namespace:kube-system,Attempt:0,} returns sandbox id \"eff336453d1400ed7af2ef59958b880a3dc233df8e893f88e67b199511d14450\"" Aug 13 07:21:44.573154 containerd[1920]: time="2025-08-13T07:21:44.573094163Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.5-a-3ca3f1276f,Uid:b57a63056e7190c205656dc6817567bc,Namespace:kube-system,Attempt:0,} returns sandbox id \"e41215942bbc1ddfc8ecc48aab4ce57be818dccd36c3f6b9741605b7aead80cf\"" Aug 13 07:21:44.575010 containerd[1920]: time="2025-08-13T07:21:44.574995732Z" level=info msg="CreateContainer within sandbox \"eff336453d1400ed7af2ef59958b880a3dc233df8e893f88e67b199511d14450\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 07:21:44.575060 containerd[1920]: time="2025-08-13T07:21:44.574997300Z" level=info msg="CreateContainer within sandbox \"e41215942bbc1ddfc8ecc48aab4ce57be818dccd36c3f6b9741605b7aead80cf\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 07:21:44.575114 containerd[1920]: time="2025-08-13T07:21:44.575060766Z" level=info msg="CreateContainer within sandbox \"05512df33da93524185c8a95df9f2d46a944cb07ea60992b1e60a7e539625732\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 07:21:44.580809 containerd[1920]: time="2025-08-13T07:21:44.580762469Z" level=info msg="CreateContainer within sandbox \"e41215942bbc1ddfc8ecc48aab4ce57be818dccd36c3f6b9741605b7aead80cf\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"a438f387c03474f319d47822b8e560b15f346a683841bd90113234f4162a12f0\"" Aug 13 07:21:44.581028 containerd[1920]: time="2025-08-13T07:21:44.581016943Z" level=info msg="StartContainer for \"a438f387c03474f319d47822b8e560b15f346a683841bd90113234f4162a12f0\"" Aug 13 07:21:44.582307 containerd[1920]: time="2025-08-13T07:21:44.582268197Z" level=info msg="CreateContainer within sandbox \"05512df33da93524185c8a95df9f2d46a944cb07ea60992b1e60a7e539625732\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"97a5835054cdee6d555a966add43ad5c807d4681a425caab1c50ba806afb49fd\"" Aug 13 07:21:44.582442 containerd[1920]: time="2025-08-13T07:21:44.582428899Z" level=info msg="StartContainer for \"97a5835054cdee6d555a966add43ad5c807d4681a425caab1c50ba806afb49fd\"" Aug 13 07:21:44.582589 containerd[1920]: time="2025-08-13T07:21:44.582576090Z" level=info msg="CreateContainer within sandbox \"eff336453d1400ed7af2ef59958b880a3dc233df8e893f88e67b199511d14450\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"84c3c589c815d0856d221c3791d9222763c57b3ac54b9669fdf906a702aab06e\"" Aug 13 07:21:44.582715 containerd[1920]: time="2025-08-13T07:21:44.582705141Z" level=info msg="StartContainer for \"84c3c589c815d0856d221c3791d9222763c57b3ac54b9669fdf906a702aab06e\"" Aug 13 07:21:44.629125 containerd[1920]: time="2025-08-13T07:21:44.629092516Z" level=info msg="StartContainer for \"97a5835054cdee6d555a966add43ad5c807d4681a425caab1c50ba806afb49fd\" returns successfully" Aug 13 07:21:44.629237 containerd[1920]: time="2025-08-13T07:21:44.629135872Z" level=info msg="StartContainer for \"a438f387c03474f319d47822b8e560b15f346a683841bd90113234f4162a12f0\" returns successfully" Aug 13 07:21:44.629237 containerd[1920]: time="2025-08-13T07:21:44.629177720Z" level=info msg="StartContainer for \"84c3c589c815d0856d221c3791d9222763c57b3ac54b9669fdf906a702aab06e\" returns successfully" Aug 13 07:21:45.094353 kubelet[2818]: I0813 07:21:45.094332 2818 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:45.303630 kubelet[2818]: E0813 07:21:45.303609 2818 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.5-a-3ca3f1276f\" not found" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:45.407718 kubelet[2818]: I0813 07:21:45.407627 2818 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:45.407718 kubelet[2818]: E0813 07:21:45.407648 2818 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4081.3.5-a-3ca3f1276f\": node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:45.411750 kubelet[2818]: E0813 07:21:45.411735 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:45.513003 kubelet[2818]: E0813 07:21:45.512928 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:45.619984 kubelet[2818]: E0813 07:21:45.619891 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:45.721178 kubelet[2818]: E0813 07:21:45.720894 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:45.821297 kubelet[2818]: E0813 07:21:45.821190 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:45.922135 kubelet[2818]: E0813 07:21:45.921998 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:46.022946 kubelet[2818]: E0813 07:21:46.022691 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:46.123633 kubelet[2818]: E0813 07:21:46.123524 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:46.224267 kubelet[2818]: E0813 07:21:46.224154 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:46.325298 kubelet[2818]: E0813 07:21:46.325197 2818 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:46.533118 kubelet[2818]: I0813 07:21:46.533042 2818 apiserver.go:52] "Watching apiserver" Aug 13 07:21:46.538561 kubelet[2818]: I0813 07:21:46.538489 2818 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 07:21:46.576537 kubelet[2818]: W0813 07:21:46.576225 2818 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:21:46.710875 kubelet[2818]: W0813 07:21:46.710809 2818 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:21:47.762435 systemd[1]: Reloading requested from client PID 3132 ('systemctl') (unit session-11.scope)... Aug 13 07:21:47.762442 systemd[1]: Reloading... Aug 13 07:21:47.804167 zram_generator::config[3171]: No configuration found. Aug 13 07:21:47.889814 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:21:47.953466 systemd[1]: Reloading finished in 190 ms. Aug 13 07:21:47.977229 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:21:47.987130 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 07:21:47.987293 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:47.998339 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:21:48.269691 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:21:48.273793 (kubelet)[3245]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:21:48.314376 kubelet[3245]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:21:48.314376 kubelet[3245]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 07:21:48.314376 kubelet[3245]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:21:48.314718 kubelet[3245]: I0813 07:21:48.314401 3245 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:21:48.320068 kubelet[3245]: I0813 07:21:48.320043 3245 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 07:21:48.320068 kubelet[3245]: I0813 07:21:48.320069 3245 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:21:48.320310 kubelet[3245]: I0813 07:21:48.320268 3245 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 07:21:48.321448 kubelet[3245]: I0813 07:21:48.321411 3245 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Aug 13 07:21:48.323021 kubelet[3245]: I0813 07:21:48.323007 3245 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:21:48.325191 kubelet[3245]: E0813 07:21:48.325169 3245 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:21:48.325191 kubelet[3245]: I0813 07:21:48.325192 3245 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:21:48.335299 kubelet[3245]: I0813 07:21:48.335259 3245 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:21:48.335593 kubelet[3245]: I0813 07:21:48.335558 3245 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 07:21:48.335696 kubelet[3245]: I0813 07:21:48.335641 3245 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:21:48.335868 kubelet[3245]: I0813 07:21:48.335663 3245 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.5-a-3ca3f1276f","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Aug 13 07:21:48.335868 kubelet[3245]: I0813 07:21:48.335843 3245 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:21:48.335868 kubelet[3245]: I0813 07:21:48.335853 3245 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 07:21:48.335868 kubelet[3245]: I0813 07:21:48.335874 3245 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:21:48.336041 kubelet[3245]: I0813 07:21:48.335942 3245 kubelet.go:408] "Attempting to sync node with API server" Aug 13 07:21:48.336041 kubelet[3245]: I0813 07:21:48.335953 3245 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:21:48.336041 kubelet[3245]: I0813 07:21:48.335975 3245 kubelet.go:314] "Adding apiserver pod source" Aug 13 07:21:48.336041 kubelet[3245]: I0813 07:21:48.335983 3245 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:21:48.336394 kubelet[3245]: I0813 07:21:48.336374 3245 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 07:21:48.336775 kubelet[3245]: I0813 07:21:48.336765 3245 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 07:21:48.337100 kubelet[3245]: I0813 07:21:48.337089 3245 server.go:1274] "Started kubelet" Aug 13 07:21:48.337209 kubelet[3245]: I0813 07:21:48.337144 3245 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:21:48.337395 kubelet[3245]: I0813 07:21:48.337151 3245 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:21:48.337623 kubelet[3245]: I0813 07:21:48.337604 3245 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:21:48.338633 kubelet[3245]: I0813 07:21:48.338613 3245 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:21:48.338710 kubelet[3245]: E0813 07:21:48.338624 3245 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:21:48.338817 kubelet[3245]: E0813 07:21:48.338791 3245 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-3ca3f1276f\" not found" Aug 13 07:21:48.338890 kubelet[3245]: I0813 07:21:48.338846 3245 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 07:21:48.338938 kubelet[3245]: I0813 07:21:48.338906 3245 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 07:21:48.338993 kubelet[3245]: I0813 07:21:48.338932 3245 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:21:48.339090 kubelet[3245]: I0813 07:21:48.339074 3245 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:21:48.340080 kubelet[3245]: I0813 07:21:48.339479 3245 factory.go:221] Registration of the systemd container factory successfully Aug 13 07:21:48.340080 kubelet[3245]: I0813 07:21:48.339690 3245 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:21:48.340080 kubelet[3245]: I0813 07:21:48.339876 3245 server.go:449] "Adding debug handlers to kubelet server" Aug 13 07:21:48.341246 kubelet[3245]: I0813 07:21:48.341229 3245 factory.go:221] Registration of the containerd container factory successfully Aug 13 07:21:48.346943 kubelet[3245]: I0813 07:21:48.346915 3245 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 07:21:48.348162 kubelet[3245]: I0813 07:21:48.348137 3245 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 07:21:48.348162 kubelet[3245]: I0813 07:21:48.348161 3245 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 07:21:48.348305 kubelet[3245]: I0813 07:21:48.348217 3245 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 07:21:48.348305 kubelet[3245]: E0813 07:21:48.348283 3245 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:21:48.371794 kubelet[3245]: I0813 07:21:48.371774 3245 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 07:21:48.371794 kubelet[3245]: I0813 07:21:48.371788 3245 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 07:21:48.371794 kubelet[3245]: I0813 07:21:48.371802 3245 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:21:48.371960 kubelet[3245]: I0813 07:21:48.371926 3245 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 07:21:48.371960 kubelet[3245]: I0813 07:21:48.371937 3245 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 07:21:48.371960 kubelet[3245]: I0813 07:21:48.371955 3245 policy_none.go:49] "None policy: Start" Aug 13 07:21:48.372426 kubelet[3245]: I0813 07:21:48.372411 3245 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 07:21:48.372466 kubelet[3245]: I0813 07:21:48.372434 3245 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:21:48.372562 kubelet[3245]: I0813 07:21:48.372552 3245 state_mem.go:75] "Updated machine memory state" Aug 13 07:21:48.373444 kubelet[3245]: I0813 07:21:48.373432 3245 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 07:21:48.373572 kubelet[3245]: I0813 07:21:48.373563 3245 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:21:48.373604 kubelet[3245]: I0813 07:21:48.373573 3245 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:21:48.373751 kubelet[3245]: I0813 07:21:48.373740 3245 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:21:48.459926 kubelet[3245]: W0813 07:21:48.459865 3245 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:21:48.460251 kubelet[3245]: W0813 07:21:48.459942 3245 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:21:48.460251 kubelet[3245]: E0813 07:21:48.460020 3245 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4081.3.5-a-3ca3f1276f\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.460251 kubelet[3245]: W0813 07:21:48.460247 3245 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:21:48.460612 kubelet[3245]: E0813 07:21:48.460341 3245 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.481310 kubelet[3245]: I0813 07:21:48.481254 3245 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.491061 kubelet[3245]: I0813 07:21:48.491019 3245 kubelet_node_status.go:111] "Node was previously registered" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.491247 kubelet[3245]: I0813 07:21:48.491183 3245 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.640338 kubelet[3245]: I0813 07:21:48.640246 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-ca-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.640338 kubelet[3245]: I0813 07:21:48.640348 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.640881 kubelet[3245]: I0813 07:21:48.640412 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.640881 kubelet[3245]: I0813 07:21:48.640466 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b2c458f081aa2526b69831ffeb469bd5-kubeconfig\") pod \"kube-scheduler-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b2c458f081aa2526b69831ffeb469bd5\") " pod="kube-system/kube-scheduler-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.640881 kubelet[3245]: I0813 07:21:48.640517 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b57a63056e7190c205656dc6817567bc-ca-certs\") pod \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b57a63056e7190c205656dc6817567bc\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.640881 kubelet[3245]: I0813 07:21:48.640564 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b57a63056e7190c205656dc6817567bc-k8s-certs\") pod \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b57a63056e7190c205656dc6817567bc\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.640881 kubelet[3245]: I0813 07:21:48.640614 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b57a63056e7190c205656dc6817567bc-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" (UID: \"b57a63056e7190c205656dc6817567bc\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.641525 kubelet[3245]: I0813 07:21:48.640669 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:48.641525 kubelet[3245]: I0813 07:21:48.640720 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/18e75a8702129840c848c326f4a5a54c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.5-a-3ca3f1276f\" (UID: \"18e75a8702129840c848c326f4a5a54c\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:49.336722 kubelet[3245]: I0813 07:21:49.336706 3245 apiserver.go:52] "Watching apiserver" Aug 13 07:21:49.339318 kubelet[3245]: I0813 07:21:49.339275 3245 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 07:21:49.356430 kubelet[3245]: W0813 07:21:49.356411 3245 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:21:49.356524 kubelet[3245]: E0813 07:21:49.356459 3245 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4081.3.5-a-3ca3f1276f\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:49.356593 kubelet[3245]: W0813 07:21:49.356581 3245 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:21:49.356632 kubelet[3245]: E0813 07:21:49.356614 3245 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081.3.5-a-3ca3f1276f\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" Aug 13 07:21:49.364261 kubelet[3245]: I0813 07:21:49.364222 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.5-a-3ca3f1276f" podStartSLOduration=3.36419747 podStartE2EDuration="3.36419747s" podCreationTimestamp="2025-08-13 07:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:21:49.364116426 +0000 UTC m=+1.087078320" watchObservedRunningTime="2025-08-13 07:21:49.36419747 +0000 UTC m=+1.087159358" Aug 13 07:21:49.371100 kubelet[3245]: I0813 07:21:49.371059 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.5-a-3ca3f1276f" podStartSLOduration=3.371044231 podStartE2EDuration="3.371044231s" podCreationTimestamp="2025-08-13 07:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:21:49.370999932 +0000 UTC m=+1.093961818" watchObservedRunningTime="2025-08-13 07:21:49.371044231 +0000 UTC m=+1.094006121" Aug 13 07:21:49.380129 kubelet[3245]: I0813 07:21:49.380099 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.5-a-3ca3f1276f" podStartSLOduration=1.380087433 podStartE2EDuration="1.380087433s" podCreationTimestamp="2025-08-13 07:21:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:21:49.376054799 +0000 UTC m=+1.099016684" watchObservedRunningTime="2025-08-13 07:21:49.380087433 +0000 UTC m=+1.103049316" Aug 13 07:21:53.245973 kubelet[3245]: I0813 07:21:53.245889 3245 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 07:21:53.247108 containerd[1920]: time="2025-08-13T07:21:53.246794234Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 07:21:53.247945 kubelet[3245]: I0813 07:21:53.247337 3245 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 07:21:53.980779 kubelet[3245]: I0813 07:21:53.980637 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169-kube-proxy\") pod \"kube-proxy-h9cwl\" (UID: \"bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169\") " pod="kube-system/kube-proxy-h9cwl" Aug 13 07:21:53.980779 kubelet[3245]: I0813 07:21:53.980751 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169-xtables-lock\") pod \"kube-proxy-h9cwl\" (UID: \"bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169\") " pod="kube-system/kube-proxy-h9cwl" Aug 13 07:21:53.981244 kubelet[3245]: I0813 07:21:53.980808 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169-lib-modules\") pod \"kube-proxy-h9cwl\" (UID: \"bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169\") " pod="kube-system/kube-proxy-h9cwl" Aug 13 07:21:53.981244 kubelet[3245]: I0813 07:21:53.980869 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpf2v\" (UniqueName: \"kubernetes.io/projected/bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169-kube-api-access-fpf2v\") pod \"kube-proxy-h9cwl\" (UID: \"bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169\") " pod="kube-system/kube-proxy-h9cwl" Aug 13 07:21:54.093641 kubelet[3245]: E0813 07:21:54.093570 3245 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Aug 13 07:21:54.093641 kubelet[3245]: E0813 07:21:54.093658 3245 projected.go:194] Error preparing data for projected volume kube-api-access-fpf2v for pod kube-system/kube-proxy-h9cwl: configmap "kube-root-ca.crt" not found Aug 13 07:21:54.094150 kubelet[3245]: E0813 07:21:54.093859 3245 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169-kube-api-access-fpf2v podName:bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169 nodeName:}" failed. No retries permitted until 2025-08-13 07:21:54.593784181 +0000 UTC m=+6.316746167 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fpf2v" (UniqueName: "kubernetes.io/projected/bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169-kube-api-access-fpf2v") pod "kube-proxy-h9cwl" (UID: "bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169") : configmap "kube-root-ca.crt" not found Aug 13 07:21:54.283131 kubelet[3245]: I0813 07:21:54.282994 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/8761a138-09ad-4869-9ad5-92dacfe3a0d4-var-lib-calico\") pod \"tigera-operator-5bf8dfcb4-zk9d8\" (UID: \"8761a138-09ad-4869-9ad5-92dacfe3a0d4\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-zk9d8" Aug 13 07:21:54.284016 kubelet[3245]: I0813 07:21:54.283151 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqf8p\" (UniqueName: \"kubernetes.io/projected/8761a138-09ad-4869-9ad5-92dacfe3a0d4-kube-api-access-xqf8p\") pod \"tigera-operator-5bf8dfcb4-zk9d8\" (UID: \"8761a138-09ad-4869-9ad5-92dacfe3a0d4\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-zk9d8" Aug 13 07:21:54.518618 containerd[1920]: time="2025-08-13T07:21:54.518486348Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-zk9d8,Uid:8761a138-09ad-4869-9ad5-92dacfe3a0d4,Namespace:tigera-operator,Attempt:0,}" Aug 13 07:21:54.530062 containerd[1920]: time="2025-08-13T07:21:54.530024091Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:21:54.530062 containerd[1920]: time="2025-08-13T07:21:54.530053059Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:21:54.530062 containerd[1920]: time="2025-08-13T07:21:54.530060086Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:54.530169 containerd[1920]: time="2025-08-13T07:21:54.530117113Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:54.580979 containerd[1920]: time="2025-08-13T07:21:54.580902570Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-zk9d8,Uid:8761a138-09ad-4869-9ad5-92dacfe3a0d4,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"e77b9ed752fb4bdee94fb6a2b1e49e66251b8201990895150943200681832e79\"" Aug 13 07:21:54.581972 containerd[1920]: time="2025-08-13T07:21:54.581955713Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Aug 13 07:21:54.805214 containerd[1920]: time="2025-08-13T07:21:54.805056864Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-h9cwl,Uid:bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169,Namespace:kube-system,Attempt:0,}" Aug 13 07:21:54.816431 containerd[1920]: time="2025-08-13T07:21:54.816264742Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:21:54.816431 containerd[1920]: time="2025-08-13T07:21:54.816362517Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:21:54.816431 containerd[1920]: time="2025-08-13T07:21:54.816372267Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:54.816568 containerd[1920]: time="2025-08-13T07:21:54.816449014Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:21:54.845636 containerd[1920]: time="2025-08-13T07:21:54.845540570Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-h9cwl,Uid:bf81f3fa-3fd6-4b9b-b45c-ac69d59cc169,Namespace:kube-system,Attempt:0,} returns sandbox id \"aa732ec4f5cff753f9d82c3b032259bd5f0616ce2b7880a03ca10d71002140d1\"" Aug 13 07:21:54.846763 containerd[1920]: time="2025-08-13T07:21:54.846747310Z" level=info msg="CreateContainer within sandbox \"aa732ec4f5cff753f9d82c3b032259bd5f0616ce2b7880a03ca10d71002140d1\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 07:21:54.851777 containerd[1920]: time="2025-08-13T07:21:54.851736628Z" level=info msg="CreateContainer within sandbox \"aa732ec4f5cff753f9d82c3b032259bd5f0616ce2b7880a03ca10d71002140d1\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"b845a0b4b7b77c55a26c264e4efb442156bb92b8dcd9cf899a7629d60e71d06f\"" Aug 13 07:21:54.851960 containerd[1920]: time="2025-08-13T07:21:54.851949371Z" level=info msg="StartContainer for \"b845a0b4b7b77c55a26c264e4efb442156bb92b8dcd9cf899a7629d60e71d06f\"" Aug 13 07:21:54.898795 containerd[1920]: time="2025-08-13T07:21:54.898743875Z" level=info msg="StartContainer for \"b845a0b4b7b77c55a26c264e4efb442156bb92b8dcd9cf899a7629d60e71d06f\" returns successfully" Aug 13 07:21:55.966324 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1817779731.mount: Deactivated successfully. Aug 13 07:21:56.329181 containerd[1920]: time="2025-08-13T07:21:56.329126900Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:56.329432 containerd[1920]: time="2025-08-13T07:21:56.329253243Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.3: active requests=0, bytes read=25056543" Aug 13 07:21:56.329658 containerd[1920]: time="2025-08-13T07:21:56.329644956Z" level=info msg="ImageCreate event name:\"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:56.330698 containerd[1920]: time="2025-08-13T07:21:56.330686255Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:21:56.331130 containerd[1920]: time="2025-08-13T07:21:56.331115697Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.3\" with image id \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\", repo tag \"quay.io/tigera/operator:v1.38.3\", repo digest \"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\", size \"25052538\" in 1.749138284s" Aug 13 07:21:56.331158 containerd[1920]: time="2025-08-13T07:21:56.331132590Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\"" Aug 13 07:21:56.332052 containerd[1920]: time="2025-08-13T07:21:56.332040186Z" level=info msg="CreateContainer within sandbox \"e77b9ed752fb4bdee94fb6a2b1e49e66251b8201990895150943200681832e79\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Aug 13 07:21:56.335473 containerd[1920]: time="2025-08-13T07:21:56.335430323Z" level=info msg="CreateContainer within sandbox \"e77b9ed752fb4bdee94fb6a2b1e49e66251b8201990895150943200681832e79\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"fe137dc112609ad53020c3b5552a1f4ad6a2277d277ce152a8ed001af78d8cbf\"" Aug 13 07:21:56.335668 containerd[1920]: time="2025-08-13T07:21:56.335652597Z" level=info msg="StartContainer for \"fe137dc112609ad53020c3b5552a1f4ad6a2277d277ce152a8ed001af78d8cbf\"" Aug 13 07:21:56.373570 containerd[1920]: time="2025-08-13T07:21:56.373548246Z" level=info msg="StartContainer for \"fe137dc112609ad53020c3b5552a1f4ad6a2277d277ce152a8ed001af78d8cbf\" returns successfully" Aug 13 07:21:57.392183 kubelet[3245]: I0813 07:21:57.392124 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-h9cwl" podStartSLOduration=4.392111985 podStartE2EDuration="4.392111985s" podCreationTimestamp="2025-08-13 07:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:21:55.377845251 +0000 UTC m=+7.100807137" watchObservedRunningTime="2025-08-13 07:21:57.392111985 +0000 UTC m=+9.115073867" Aug 13 07:21:57.392460 kubelet[3245]: I0813 07:21:57.392199 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-5bf8dfcb4-zk9d8" podStartSLOduration=1.642406771 podStartE2EDuration="3.392195862s" podCreationTimestamp="2025-08-13 07:21:54 +0000 UTC" firstStartedPulling="2025-08-13 07:21:54.581706997 +0000 UTC m=+6.304668888" lastFinishedPulling="2025-08-13 07:21:56.331496093 +0000 UTC m=+8.054457979" observedRunningTime="2025-08-13 07:21:57.392020655 +0000 UTC m=+9.114982542" watchObservedRunningTime="2025-08-13 07:21:57.392195862 +0000 UTC m=+9.115157744" Aug 13 07:22:00.739109 sudo[2245]: pam_unix(sudo:session): session closed for user root Aug 13 07:22:00.740003 sshd[2241]: pam_unix(sshd:session): session closed for user core Aug 13 07:22:00.743370 systemd[1]: sshd@8-145.40.90.243:22-147.75.109.163:36212.service: Deactivated successfully. Aug 13 07:22:00.744501 systemd-logind[1908]: Session 11 logged out. Waiting for processes to exit. Aug 13 07:22:00.744503 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 07:22:00.745018 systemd-logind[1908]: Removed session 11. Aug 13 07:22:02.104204 update_engine[1913]: I20250813 07:22:02.104128 1913 update_attempter.cc:509] Updating boot flags... Aug 13 07:22:02.136077 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 37 scanned by (udev-worker) (3772) Aug 13 07:22:02.164076 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 37 scanned by (udev-worker) (3772) Aug 13 07:22:02.184076 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 37 scanned by (udev-worker) (3772) Aug 13 07:22:04.553258 kubelet[3245]: I0813 07:22:04.553201 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/de4bfe78-d274-4922-a241-06b61c5b700b-typha-certs\") pod \"calico-typha-64dcb5988-fbcfp\" (UID: \"de4bfe78-d274-4922-a241-06b61c5b700b\") " pod="calico-system/calico-typha-64dcb5988-fbcfp" Aug 13 07:22:04.553258 kubelet[3245]: I0813 07:22:04.553229 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw4s9\" (UniqueName: \"kubernetes.io/projected/de4bfe78-d274-4922-a241-06b61c5b700b-kube-api-access-lw4s9\") pod \"calico-typha-64dcb5988-fbcfp\" (UID: \"de4bfe78-d274-4922-a241-06b61c5b700b\") " pod="calico-system/calico-typha-64dcb5988-fbcfp" Aug 13 07:22:04.553258 kubelet[3245]: I0813 07:22:04.553245 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/de4bfe78-d274-4922-a241-06b61c5b700b-tigera-ca-bundle\") pod \"calico-typha-64dcb5988-fbcfp\" (UID: \"de4bfe78-d274-4922-a241-06b61c5b700b\") " pod="calico-system/calico-typha-64dcb5988-fbcfp" Aug 13 07:22:04.816148 containerd[1920]: time="2025-08-13T07:22:04.815936029Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-64dcb5988-fbcfp,Uid:de4bfe78-d274-4922-a241-06b61c5b700b,Namespace:calico-system,Attempt:0,}" Aug 13 07:22:04.828275 containerd[1920]: time="2025-08-13T07:22:04.828228607Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:04.828275 containerd[1920]: time="2025-08-13T07:22:04.828263036Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:04.828275 containerd[1920]: time="2025-08-13T07:22:04.828270524Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:04.828441 containerd[1920]: time="2025-08-13T07:22:04.828322075Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:04.855625 kubelet[3245]: I0813 07:22:04.855576 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-policysync\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855625 kubelet[3245]: I0813 07:22:04.855604 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-var-run-calico\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855625 kubelet[3245]: I0813 07:22:04.855621 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-cni-bin-dir\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855741 kubelet[3245]: I0813 07:22:04.855637 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-lib-modules\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855741 kubelet[3245]: I0813 07:22:04.855653 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpw7z\" (UniqueName: \"kubernetes.io/projected/922e4266-8622-4cf8-86b4-376e98d45e89-kube-api-access-tpw7z\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855741 kubelet[3245]: I0813 07:22:04.855672 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-xtables-lock\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855741 kubelet[3245]: I0813 07:22:04.855689 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-cni-log-dir\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855741 kubelet[3245]: I0813 07:22:04.855704 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-flexvol-driver-host\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855829 kubelet[3245]: I0813 07:22:04.855720 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-cni-net-dir\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855829 kubelet[3245]: I0813 07:22:04.855737 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/922e4266-8622-4cf8-86b4-376e98d45e89-node-certs\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855829 kubelet[3245]: I0813 07:22:04.855757 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/922e4266-8622-4cf8-86b4-376e98d45e89-tigera-ca-bundle\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.855829 kubelet[3245]: I0813 07:22:04.855777 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/922e4266-8622-4cf8-86b4-376e98d45e89-var-lib-calico\") pod \"calico-node-s8brb\" (UID: \"922e4266-8622-4cf8-86b4-376e98d45e89\") " pod="calico-system/calico-node-s8brb" Aug 13 07:22:04.866588 containerd[1920]: time="2025-08-13T07:22:04.866565601Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-64dcb5988-fbcfp,Uid:de4bfe78-d274-4922-a241-06b61c5b700b,Namespace:calico-system,Attempt:0,} returns sandbox id \"e03b90c17df2fb33ca5c29527a525b19642bfb7dab4f308d92452191fe6cd1db\"" Aug 13 07:22:04.867272 containerd[1920]: time="2025-08-13T07:22:04.867259966Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Aug 13 07:22:04.959686 kubelet[3245]: E0813 07:22:04.959591 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:04.959686 kubelet[3245]: W0813 07:22:04.959642 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:04.959686 kubelet[3245]: E0813 07:22:04.959692 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:04.965810 kubelet[3245]: E0813 07:22:04.965762 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:04.965810 kubelet[3245]: W0813 07:22:04.965805 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:04.966237 kubelet[3245]: E0813 07:22:04.965847 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:04.977656 kubelet[3245]: E0813 07:22:04.977577 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:04.977656 kubelet[3245]: W0813 07:22:04.977616 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:04.977656 kubelet[3245]: E0813 07:22:04.977653 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.099381 kubelet[3245]: E0813 07:22:05.099113 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xnn8d" podUID="b90d45be-b93c-47a4-ad3d-74de560838fa" Aug 13 07:22:05.133272 containerd[1920]: time="2025-08-13T07:22:05.133249431Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-s8brb,Uid:922e4266-8622-4cf8-86b4-376e98d45e89,Namespace:calico-system,Attempt:0,}" Aug 13 07:22:05.142660 containerd[1920]: time="2025-08-13T07:22:05.142407945Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:05.142660 containerd[1920]: time="2025-08-13T07:22:05.142624065Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:05.142660 containerd[1920]: time="2025-08-13T07:22:05.142631430Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:05.142816 containerd[1920]: time="2025-08-13T07:22:05.142680211Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:05.151763 kubelet[3245]: E0813 07:22:05.151718 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.151763 kubelet[3245]: W0813 07:22:05.151733 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.151763 kubelet[3245]: E0813 07:22:05.151748 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.151906 kubelet[3245]: E0813 07:22:05.151882 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.151906 kubelet[3245]: W0813 07:22:05.151887 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.151906 kubelet[3245]: E0813 07:22:05.151893 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.151971 kubelet[3245]: E0813 07:22:05.151966 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.151991 kubelet[3245]: W0813 07:22:05.151971 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.151991 kubelet[3245]: E0813 07:22:05.151976 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152077 kubelet[3245]: E0813 07:22:05.152063 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152077 kubelet[3245]: W0813 07:22:05.152073 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152077 kubelet[3245]: E0813 07:22:05.152078 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152239 kubelet[3245]: E0813 07:22:05.152204 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152239 kubelet[3245]: W0813 07:22:05.152211 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152239 kubelet[3245]: E0813 07:22:05.152217 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152363 kubelet[3245]: E0813 07:22:05.152328 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152363 kubelet[3245]: W0813 07:22:05.152334 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152363 kubelet[3245]: E0813 07:22:05.152341 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152469 kubelet[3245]: E0813 07:22:05.152464 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152487 kubelet[3245]: W0813 07:22:05.152469 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152487 kubelet[3245]: E0813 07:22:05.152474 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152555 kubelet[3245]: E0813 07:22:05.152550 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152574 kubelet[3245]: W0813 07:22:05.152555 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152574 kubelet[3245]: E0813 07:22:05.152560 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152678 kubelet[3245]: E0813 07:22:05.152673 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152697 kubelet[3245]: W0813 07:22:05.152678 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152697 kubelet[3245]: E0813 07:22:05.152684 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152789 kubelet[3245]: E0813 07:22:05.152784 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152809 kubelet[3245]: W0813 07:22:05.152789 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152809 kubelet[3245]: E0813 07:22:05.152793 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152890 kubelet[3245]: E0813 07:22:05.152886 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.152910 kubelet[3245]: W0813 07:22:05.152890 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.152910 kubelet[3245]: E0813 07:22:05.152894 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.152989 kubelet[3245]: E0813 07:22:05.152984 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153008 kubelet[3245]: W0813 07:22:05.152989 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153008 kubelet[3245]: E0813 07:22:05.152993 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153072 kubelet[3245]: E0813 07:22:05.153059 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153072 kubelet[3245]: W0813 07:22:05.153068 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153109 kubelet[3245]: E0813 07:22:05.153077 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153151 kubelet[3245]: E0813 07:22:05.153146 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153168 kubelet[3245]: W0813 07:22:05.153151 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153168 kubelet[3245]: E0813 07:22:05.153155 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153259 kubelet[3245]: E0813 07:22:05.153255 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153259 kubelet[3245]: W0813 07:22:05.153259 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153292 kubelet[3245]: E0813 07:22:05.153263 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153369 kubelet[3245]: E0813 07:22:05.153364 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153387 kubelet[3245]: W0813 07:22:05.153369 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153387 kubelet[3245]: E0813 07:22:05.153373 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153478 kubelet[3245]: E0813 07:22:05.153473 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153497 kubelet[3245]: W0813 07:22:05.153478 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153497 kubelet[3245]: E0813 07:22:05.153483 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153577 kubelet[3245]: E0813 07:22:05.153573 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153594 kubelet[3245]: W0813 07:22:05.153577 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153594 kubelet[3245]: E0813 07:22:05.153582 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153648 kubelet[3245]: E0813 07:22:05.153644 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153668 kubelet[3245]: W0813 07:22:05.153648 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153668 kubelet[3245]: E0813 07:22:05.153653 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.153721 kubelet[3245]: E0813 07:22:05.153717 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.153740 kubelet[3245]: W0813 07:22:05.153722 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.153740 kubelet[3245]: E0813 07:22:05.153726 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158240 kubelet[3245]: E0813 07:22:05.158232 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.158240 kubelet[3245]: W0813 07:22:05.158239 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.158299 kubelet[3245]: E0813 07:22:05.158246 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158299 kubelet[3245]: I0813 07:22:05.158260 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b90d45be-b93c-47a4-ad3d-74de560838fa-registration-dir\") pod \"csi-node-driver-xnn8d\" (UID: \"b90d45be-b93c-47a4-ad3d-74de560838fa\") " pod="calico-system/csi-node-driver-xnn8d" Aug 13 07:22:05.158351 kubelet[3245]: E0813 07:22:05.158344 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.158369 kubelet[3245]: W0813 07:22:05.158350 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.158369 kubelet[3245]: E0813 07:22:05.158357 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158369 kubelet[3245]: I0813 07:22:05.158365 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b90d45be-b93c-47a4-ad3d-74de560838fa-socket-dir\") pod \"csi-node-driver-xnn8d\" (UID: \"b90d45be-b93c-47a4-ad3d-74de560838fa\") " pod="calico-system/csi-node-driver-xnn8d" Aug 13 07:22:05.158486 kubelet[3245]: E0813 07:22:05.158481 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.158504 kubelet[3245]: W0813 07:22:05.158486 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.158504 kubelet[3245]: E0813 07:22:05.158493 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158504 kubelet[3245]: I0813 07:22:05.158500 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9cnt\" (UniqueName: \"kubernetes.io/projected/b90d45be-b93c-47a4-ad3d-74de560838fa-kube-api-access-q9cnt\") pod \"csi-node-driver-xnn8d\" (UID: \"b90d45be-b93c-47a4-ad3d-74de560838fa\") " pod="calico-system/csi-node-driver-xnn8d" Aug 13 07:22:05.158620 kubelet[3245]: E0813 07:22:05.158615 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.158642 kubelet[3245]: W0813 07:22:05.158620 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.158642 kubelet[3245]: E0813 07:22:05.158627 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158642 kubelet[3245]: I0813 07:22:05.158635 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b90d45be-b93c-47a4-ad3d-74de560838fa-kubelet-dir\") pod \"csi-node-driver-xnn8d\" (UID: \"b90d45be-b93c-47a4-ad3d-74de560838fa\") " pod="calico-system/csi-node-driver-xnn8d" Aug 13 07:22:05.158735 kubelet[3245]: E0813 07:22:05.158730 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.158754 kubelet[3245]: W0813 07:22:05.158735 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.158754 kubelet[3245]: E0813 07:22:05.158741 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158754 kubelet[3245]: I0813 07:22:05.158748 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/b90d45be-b93c-47a4-ad3d-74de560838fa-varrun\") pod \"csi-node-driver-xnn8d\" (UID: \"b90d45be-b93c-47a4-ad3d-74de560838fa\") " pod="calico-system/csi-node-driver-xnn8d" Aug 13 07:22:05.158836 kubelet[3245]: E0813 07:22:05.158831 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.158853 kubelet[3245]: W0813 07:22:05.158836 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.158853 kubelet[3245]: E0813 07:22:05.158842 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158912 kubelet[3245]: E0813 07:22:05.158908 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.158933 kubelet[3245]: W0813 07:22:05.158912 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.158933 kubelet[3245]: E0813 07:22:05.158918 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.158998 kubelet[3245]: E0813 07:22:05.158993 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159014 kubelet[3245]: W0813 07:22:05.158998 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159014 kubelet[3245]: E0813 07:22:05.159003 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.159079 kubelet[3245]: E0813 07:22:05.159074 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159079 kubelet[3245]: W0813 07:22:05.159079 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159112 kubelet[3245]: E0813 07:22:05.159084 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.159162 kubelet[3245]: E0813 07:22:05.159158 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159179 kubelet[3245]: W0813 07:22:05.159162 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159179 kubelet[3245]: E0813 07:22:05.159168 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.159237 kubelet[3245]: E0813 07:22:05.159233 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159257 kubelet[3245]: W0813 07:22:05.159237 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159257 kubelet[3245]: E0813 07:22:05.159243 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.159319 kubelet[3245]: E0813 07:22:05.159314 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159336 kubelet[3245]: W0813 07:22:05.159319 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159336 kubelet[3245]: E0813 07:22:05.159325 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.159393 kubelet[3245]: E0813 07:22:05.159389 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159415 kubelet[3245]: W0813 07:22:05.159394 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159415 kubelet[3245]: E0813 07:22:05.159399 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.159473 kubelet[3245]: E0813 07:22:05.159469 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159492 kubelet[3245]: W0813 07:22:05.159473 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159492 kubelet[3245]: E0813 07:22:05.159478 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.159546 kubelet[3245]: E0813 07:22:05.159541 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.159565 kubelet[3245]: W0813 07:22:05.159546 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.159565 kubelet[3245]: E0813 07:22:05.159550 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.165037 containerd[1920]: time="2025-08-13T07:22:05.165013383Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-s8brb,Uid:922e4266-8622-4cf8-86b4-376e98d45e89,Namespace:calico-system,Attempt:0,} returns sandbox id \"3410d01fc2075fca23a48d575638f24f9bed57f49e04f881b51dbe1f880902ae\"" Aug 13 07:22:05.259865 kubelet[3245]: E0813 07:22:05.259810 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.259865 kubelet[3245]: W0813 07:22:05.259860 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.260305 kubelet[3245]: E0813 07:22:05.259905 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.260543 kubelet[3245]: E0813 07:22:05.260496 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.260543 kubelet[3245]: W0813 07:22:05.260534 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.260887 kubelet[3245]: E0813 07:22:05.260578 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.261363 kubelet[3245]: E0813 07:22:05.261276 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.261363 kubelet[3245]: W0813 07:22:05.261326 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.261715 kubelet[3245]: E0813 07:22:05.261388 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.261924 kubelet[3245]: E0813 07:22:05.261867 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.261924 kubelet[3245]: W0813 07:22:05.261903 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.262251 kubelet[3245]: E0813 07:22:05.261943 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.262452 kubelet[3245]: E0813 07:22:05.262423 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.262566 kubelet[3245]: W0813 07:22:05.262454 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.262709 kubelet[3245]: E0813 07:22:05.262560 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.262993 kubelet[3245]: E0813 07:22:05.262965 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.263148 kubelet[3245]: W0813 07:22:05.262994 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.263148 kubelet[3245]: E0813 07:22:05.263102 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.263538 kubelet[3245]: E0813 07:22:05.263504 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.263724 kubelet[3245]: W0813 07:22:05.263537 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.263724 kubelet[3245]: E0813 07:22:05.263626 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.264202 kubelet[3245]: E0813 07:22:05.264149 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.264202 kubelet[3245]: W0813 07:22:05.264177 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.264565 kubelet[3245]: E0813 07:22:05.264213 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.264867 kubelet[3245]: E0813 07:22:05.264822 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.264867 kubelet[3245]: W0813 07:22:05.264863 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.265231 kubelet[3245]: E0813 07:22:05.264921 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.265554 kubelet[3245]: E0813 07:22:05.265498 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.265554 kubelet[3245]: W0813 07:22:05.265535 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.265900 kubelet[3245]: E0813 07:22:05.265581 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.266299 kubelet[3245]: E0813 07:22:05.266253 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.266299 kubelet[3245]: W0813 07:22:05.266294 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.266585 kubelet[3245]: E0813 07:22:05.266350 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.266953 kubelet[3245]: E0813 07:22:05.266905 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.266953 kubelet[3245]: W0813 07:22:05.266945 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.267296 kubelet[3245]: E0813 07:22:05.266992 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.267731 kubelet[3245]: E0813 07:22:05.267651 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.267731 kubelet[3245]: W0813 07:22:05.267694 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.267980 kubelet[3245]: E0813 07:22:05.267740 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.268335 kubelet[3245]: E0813 07:22:05.268277 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.268335 kubelet[3245]: W0813 07:22:05.268307 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.268570 kubelet[3245]: E0813 07:22:05.268392 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.268859 kubelet[3245]: E0813 07:22:05.268826 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.268859 kubelet[3245]: W0813 07:22:05.268853 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.269139 kubelet[3245]: E0813 07:22:05.268910 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.269480 kubelet[3245]: E0813 07:22:05.269413 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.269480 kubelet[3245]: W0813 07:22:05.269456 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.269729 kubelet[3245]: E0813 07:22:05.269587 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.270038 kubelet[3245]: E0813 07:22:05.270007 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.270038 kubelet[3245]: W0813 07:22:05.270038 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.270414 kubelet[3245]: E0813 07:22:05.270143 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.270683 kubelet[3245]: E0813 07:22:05.270647 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.270683 kubelet[3245]: W0813 07:22:05.270677 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.270917 kubelet[3245]: E0813 07:22:05.270788 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.271200 kubelet[3245]: E0813 07:22:05.271153 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.271200 kubelet[3245]: W0813 07:22:05.271179 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.271456 kubelet[3245]: E0813 07:22:05.271282 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.271755 kubelet[3245]: E0813 07:22:05.271702 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.271755 kubelet[3245]: W0813 07:22:05.271729 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.271978 kubelet[3245]: E0813 07:22:05.271832 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.272253 kubelet[3245]: E0813 07:22:05.272187 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.272253 kubelet[3245]: W0813 07:22:05.272213 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.272501 kubelet[3245]: E0813 07:22:05.272317 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.272815 kubelet[3245]: E0813 07:22:05.272767 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.272815 kubelet[3245]: W0813 07:22:05.272794 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.273050 kubelet[3245]: E0813 07:22:05.272887 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.273379 kubelet[3245]: E0813 07:22:05.273331 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.273379 kubelet[3245]: W0813 07:22:05.273359 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.273624 kubelet[3245]: E0813 07:22:05.273445 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.273878 kubelet[3245]: E0813 07:22:05.273818 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.273878 kubelet[3245]: W0813 07:22:05.273846 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.273878 kubelet[3245]: E0813 07:22:05.273875 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.274510 kubelet[3245]: E0813 07:22:05.274457 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.274510 kubelet[3245]: W0813 07:22:05.274485 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.274510 kubelet[3245]: E0813 07:22:05.274514 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:05.296267 kubelet[3245]: E0813 07:22:05.296179 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:05.296267 kubelet[3245]: W0813 07:22:05.296225 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:05.296267 kubelet[3245]: E0813 07:22:05.296266 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:06.405288 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3679676547.mount: Deactivated successfully. Aug 13 07:22:06.707097 containerd[1920]: time="2025-08-13T07:22:06.707034624Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:06.707302 containerd[1920]: time="2025-08-13T07:22:06.707241206Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.2: active requests=0, bytes read=35233364" Aug 13 07:22:06.707649 containerd[1920]: time="2025-08-13T07:22:06.707607945Z" level=info msg="ImageCreate event name:\"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:06.708669 containerd[1920]: time="2025-08-13T07:22:06.708629136Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:06.709369 containerd[1920]: time="2025-08-13T07:22:06.709326610Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.2\" with image id \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\", size \"35233218\" in 1.842049596s" Aug 13 07:22:06.709369 containerd[1920]: time="2025-08-13T07:22:06.709343144Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\"" Aug 13 07:22:06.709765 containerd[1920]: time="2025-08-13T07:22:06.709723277Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Aug 13 07:22:06.712614 containerd[1920]: time="2025-08-13T07:22:06.712573724Z" level=info msg="CreateContainer within sandbox \"e03b90c17df2fb33ca5c29527a525b19642bfb7dab4f308d92452191fe6cd1db\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Aug 13 07:22:06.716452 containerd[1920]: time="2025-08-13T07:22:06.716435492Z" level=info msg="CreateContainer within sandbox \"e03b90c17df2fb33ca5c29527a525b19642bfb7dab4f308d92452191fe6cd1db\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"3006de67e6ca1f33c05c70c945ece90a3c23f98f7f4610eb83e39684a27ce0e5\"" Aug 13 07:22:06.716702 containerd[1920]: time="2025-08-13T07:22:06.716690583Z" level=info msg="StartContainer for \"3006de67e6ca1f33c05c70c945ece90a3c23f98f7f4610eb83e39684a27ce0e5\"" Aug 13 07:22:06.771990 containerd[1920]: time="2025-08-13T07:22:06.771966939Z" level=info msg="StartContainer for \"3006de67e6ca1f33c05c70c945ece90a3c23f98f7f4610eb83e39684a27ce0e5\" returns successfully" Aug 13 07:22:07.348952 kubelet[3245]: E0813 07:22:07.348853 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xnn8d" podUID="b90d45be-b93c-47a4-ad3d-74de560838fa" Aug 13 07:22:07.423900 kubelet[3245]: I0813 07:22:07.423790 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-64dcb5988-fbcfp" podStartSLOduration=1.581222723 podStartE2EDuration="3.423750265s" podCreationTimestamp="2025-08-13 07:22:04 +0000 UTC" firstStartedPulling="2025-08-13 07:22:04.867132073 +0000 UTC m=+16.590093961" lastFinishedPulling="2025-08-13 07:22:06.709659618 +0000 UTC m=+18.432621503" observedRunningTime="2025-08-13 07:22:07.423233021 +0000 UTC m=+19.146194976" watchObservedRunningTime="2025-08-13 07:22:07.423750265 +0000 UTC m=+19.146712203" Aug 13 07:22:07.472089 kubelet[3245]: E0813 07:22:07.472020 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.472354 kubelet[3245]: W0813 07:22:07.472097 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.472354 kubelet[3245]: E0813 07:22:07.472152 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.472865 kubelet[3245]: E0813 07:22:07.472815 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.472865 kubelet[3245]: W0813 07:22:07.472846 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.473105 kubelet[3245]: E0813 07:22:07.472878 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.473683 kubelet[3245]: E0813 07:22:07.473621 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.473683 kubelet[3245]: W0813 07:22:07.473652 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.473683 kubelet[3245]: E0813 07:22:07.473683 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.474311 kubelet[3245]: E0813 07:22:07.474273 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.474311 kubelet[3245]: W0813 07:22:07.474304 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.474669 kubelet[3245]: E0813 07:22:07.474334 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.474910 kubelet[3245]: E0813 07:22:07.474877 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.475019 kubelet[3245]: W0813 07:22:07.474911 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.475019 kubelet[3245]: E0813 07:22:07.474945 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.475515 kubelet[3245]: E0813 07:22:07.475485 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.475624 kubelet[3245]: W0813 07:22:07.475515 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.475624 kubelet[3245]: E0813 07:22:07.475545 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.476013 kubelet[3245]: E0813 07:22:07.475986 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.476157 kubelet[3245]: W0813 07:22:07.476016 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.476157 kubelet[3245]: E0813 07:22:07.476044 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.476674 kubelet[3245]: E0813 07:22:07.476623 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.476674 kubelet[3245]: W0813 07:22:07.476652 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.476890 kubelet[3245]: E0813 07:22:07.476680 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.477354 kubelet[3245]: E0813 07:22:07.477292 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.477354 kubelet[3245]: W0813 07:22:07.477321 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.477354 kubelet[3245]: E0813 07:22:07.477350 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.477914 kubelet[3245]: E0813 07:22:07.477871 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.478128 kubelet[3245]: W0813 07:22:07.477914 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.478128 kubelet[3245]: E0813 07:22:07.477958 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.478530 kubelet[3245]: E0813 07:22:07.478489 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.478530 kubelet[3245]: W0813 07:22:07.478517 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.478772 kubelet[3245]: E0813 07:22:07.478550 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.479043 kubelet[3245]: E0813 07:22:07.479014 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.479188 kubelet[3245]: W0813 07:22:07.479043 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.479188 kubelet[3245]: E0813 07:22:07.479093 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.479571 kubelet[3245]: E0813 07:22:07.479524 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.479571 kubelet[3245]: W0813 07:22:07.479551 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.479795 kubelet[3245]: E0813 07:22:07.479578 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.480135 kubelet[3245]: E0813 07:22:07.480097 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.480135 kubelet[3245]: W0813 07:22:07.480127 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.480553 kubelet[3245]: E0813 07:22:07.480154 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.480747 kubelet[3245]: E0813 07:22:07.480664 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.480747 kubelet[3245]: W0813 07:22:07.480690 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.480747 kubelet[3245]: E0813 07:22:07.480716 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.484300 kubelet[3245]: E0813 07:22:07.484227 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.484300 kubelet[3245]: W0813 07:22:07.484264 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.484300 kubelet[3245]: E0813 07:22:07.484301 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.484902 kubelet[3245]: E0813 07:22:07.484845 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.484902 kubelet[3245]: W0813 07:22:07.484876 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.485234 kubelet[3245]: E0813 07:22:07.484915 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.485596 kubelet[3245]: E0813 07:22:07.485505 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.485596 kubelet[3245]: W0813 07:22:07.485554 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.485857 kubelet[3245]: E0813 07:22:07.485603 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.486227 kubelet[3245]: E0813 07:22:07.486156 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.486227 kubelet[3245]: W0813 07:22:07.486186 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.486227 kubelet[3245]: E0813 07:22:07.486222 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.486832 kubelet[3245]: E0813 07:22:07.486772 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.486832 kubelet[3245]: W0813 07:22:07.486810 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.487124 kubelet[3245]: E0813 07:22:07.486905 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.487481 kubelet[3245]: E0813 07:22:07.487442 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.487481 kubelet[3245]: W0813 07:22:07.487479 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.487754 kubelet[3245]: E0813 07:22:07.487570 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.488008 kubelet[3245]: E0813 07:22:07.487975 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.488008 kubelet[3245]: W0813 07:22:07.488006 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.488275 kubelet[3245]: E0813 07:22:07.488107 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.488631 kubelet[3245]: E0813 07:22:07.488597 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.488631 kubelet[3245]: W0813 07:22:07.488626 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.488890 kubelet[3245]: E0813 07:22:07.488714 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.489206 kubelet[3245]: E0813 07:22:07.489166 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.489206 kubelet[3245]: W0813 07:22:07.489194 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.489438 kubelet[3245]: E0813 07:22:07.489229 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.489789 kubelet[3245]: E0813 07:22:07.489758 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.489976 kubelet[3245]: W0813 07:22:07.489787 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.489976 kubelet[3245]: E0813 07:22:07.489859 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.490267 kubelet[3245]: E0813 07:22:07.490220 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.490267 kubelet[3245]: W0813 07:22:07.490250 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.490536 kubelet[3245]: E0813 07:22:07.490319 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.490777 kubelet[3245]: E0813 07:22:07.490729 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.490777 kubelet[3245]: W0813 07:22:07.490755 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.491081 kubelet[3245]: E0813 07:22:07.490824 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.491349 kubelet[3245]: E0813 07:22:07.491312 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.491349 kubelet[3245]: W0813 07:22:07.491341 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.491656 kubelet[3245]: E0813 07:22:07.491377 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.491973 kubelet[3245]: E0813 07:22:07.491938 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.491973 kubelet[3245]: W0813 07:22:07.491967 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.492333 kubelet[3245]: E0813 07:22:07.492093 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.492543 kubelet[3245]: E0813 07:22:07.492509 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.492543 kubelet[3245]: W0813 07:22:07.492536 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.492845 kubelet[3245]: E0813 07:22:07.492565 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.493175 kubelet[3245]: E0813 07:22:07.493142 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.493175 kubelet[3245]: W0813 07:22:07.493170 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.493405 kubelet[3245]: E0813 07:22:07.493203 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.493948 kubelet[3245]: E0813 07:22:07.493914 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.494049 kubelet[3245]: W0813 07:22:07.493953 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.494049 kubelet[3245]: E0813 07:22:07.493999 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:07.494643 kubelet[3245]: E0813 07:22:07.494608 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:07.494745 kubelet[3245]: W0813 07:22:07.494647 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:07.494745 kubelet[3245]: E0813 07:22:07.494683 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.340431 containerd[1920]: time="2025-08-13T07:22:08.340405706Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:08.341113 containerd[1920]: time="2025-08-13T07:22:08.341057452Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2: active requests=0, bytes read=4446956" Aug 13 07:22:08.341794 containerd[1920]: time="2025-08-13T07:22:08.341779690Z" level=info msg="ImageCreate event name:\"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:08.343122 containerd[1920]: time="2025-08-13T07:22:08.343093376Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:08.343388 containerd[1920]: time="2025-08-13T07:22:08.343374062Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" with image id \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\", size \"5939619\" in 1.633636475s" Aug 13 07:22:08.343424 containerd[1920]: time="2025-08-13T07:22:08.343389592Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\"" Aug 13 07:22:08.344452 containerd[1920]: time="2025-08-13T07:22:08.344440195Z" level=info msg="CreateContainer within sandbox \"3410d01fc2075fca23a48d575638f24f9bed57f49e04f881b51dbe1f880902ae\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Aug 13 07:22:08.402172 kubelet[3245]: I0813 07:22:08.402143 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:08.488045 kubelet[3245]: E0813 07:22:08.488012 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.488045 kubelet[3245]: W0813 07:22:08.488036 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.488279 kubelet[3245]: E0813 07:22:08.488077 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.488382 kubelet[3245]: E0813 07:22:08.488363 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.488466 kubelet[3245]: W0813 07:22:08.488382 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.488466 kubelet[3245]: E0813 07:22:08.488404 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.488744 kubelet[3245]: E0813 07:22:08.488718 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.488744 kubelet[3245]: W0813 07:22:08.488739 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.488921 kubelet[3245]: E0813 07:22:08.488763 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.489096 kubelet[3245]: E0813 07:22:08.489074 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.489096 kubelet[3245]: W0813 07:22:08.489091 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.489297 kubelet[3245]: E0813 07:22:08.489111 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.489451 kubelet[3245]: E0813 07:22:08.489434 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.489451 kubelet[3245]: W0813 07:22:08.489448 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.489585 kubelet[3245]: E0813 07:22:08.489467 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.489705 kubelet[3245]: E0813 07:22:08.489676 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.489705 kubelet[3245]: W0813 07:22:08.489691 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.489827 kubelet[3245]: E0813 07:22:08.489712 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.489916 kubelet[3245]: E0813 07:22:08.489903 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.489987 kubelet[3245]: W0813 07:22:08.489916 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.489987 kubelet[3245]: E0813 07:22:08.489932 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.490157 kubelet[3245]: E0813 07:22:08.490143 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.490240 kubelet[3245]: W0813 07:22:08.490157 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.490240 kubelet[3245]: E0813 07:22:08.490173 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.490391 kubelet[3245]: E0813 07:22:08.490378 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.490463 kubelet[3245]: W0813 07:22:08.490391 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.490463 kubelet[3245]: E0813 07:22:08.490407 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.490605 kubelet[3245]: E0813 07:22:08.490592 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.490605 kubelet[3245]: W0813 07:22:08.490605 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.490730 kubelet[3245]: E0813 07:22:08.490620 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.490813 kubelet[3245]: E0813 07:22:08.490801 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.490883 kubelet[3245]: W0813 07:22:08.490813 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.490883 kubelet[3245]: E0813 07:22:08.490829 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.491027 kubelet[3245]: E0813 07:22:08.491014 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.491108 kubelet[3245]: W0813 07:22:08.491026 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.491108 kubelet[3245]: E0813 07:22:08.491042 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.491279 kubelet[3245]: E0813 07:22:08.491266 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.491279 kubelet[3245]: W0813 07:22:08.491278 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.491413 kubelet[3245]: E0813 07:22:08.491294 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.491512 kubelet[3245]: E0813 07:22:08.491499 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.491583 kubelet[3245]: W0813 07:22:08.491512 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.491583 kubelet[3245]: E0813 07:22:08.491528 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.491735 kubelet[3245]: E0813 07:22:08.491722 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.491807 kubelet[3245]: W0813 07:22:08.491735 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.491807 kubelet[3245]: E0813 07:22:08.491750 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.496610 containerd[1920]: time="2025-08-13T07:22:08.496575205Z" level=info msg="CreateContainer within sandbox \"3410d01fc2075fca23a48d575638f24f9bed57f49e04f881b51dbe1f880902ae\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"162dd8242386a801ed86b3511cb02e64723c3dc63349420bb3fdd3e301ebd1b7\"" Aug 13 07:22:08.496984 containerd[1920]: time="2025-08-13T07:22:08.496958725Z" level=info msg="StartContainer for \"162dd8242386a801ed86b3511cb02e64723c3dc63349420bb3fdd3e301ebd1b7\"" Aug 13 07:22:08.497150 kubelet[3245]: E0813 07:22:08.497133 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.497204 kubelet[3245]: W0813 07:22:08.497153 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.497204 kubelet[3245]: E0813 07:22:08.497177 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.497503 kubelet[3245]: E0813 07:22:08.497487 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.497554 kubelet[3245]: W0813 07:22:08.497505 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.497554 kubelet[3245]: E0813 07:22:08.497528 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.497837 kubelet[3245]: E0813 07:22:08.497820 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.497837 kubelet[3245]: W0813 07:22:08.497836 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.497923 kubelet[3245]: E0813 07:22:08.497854 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.498101 kubelet[3245]: E0813 07:22:08.498088 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.498101 kubelet[3245]: W0813 07:22:08.498100 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.498194 kubelet[3245]: E0813 07:22:08.498114 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.498320 kubelet[3245]: E0813 07:22:08.498307 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.498373 kubelet[3245]: W0813 07:22:08.498323 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.498373 kubelet[3245]: E0813 07:22:08.498342 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.498651 kubelet[3245]: E0813 07:22:08.498640 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.498703 kubelet[3245]: W0813 07:22:08.498652 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.498703 kubelet[3245]: E0813 07:22:08.498668 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.499000 kubelet[3245]: E0813 07:22:08.498983 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.499089 kubelet[3245]: W0813 07:22:08.499000 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.499089 kubelet[3245]: E0813 07:22:08.499024 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.499315 kubelet[3245]: E0813 07:22:08.499301 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.499391 kubelet[3245]: W0813 07:22:08.499315 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.499391 kubelet[3245]: E0813 07:22:08.499348 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.499543 kubelet[3245]: E0813 07:22:08.499529 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.499543 kubelet[3245]: W0813 07:22:08.499543 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.499669 kubelet[3245]: E0813 07:22:08.499568 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.499754 kubelet[3245]: E0813 07:22:08.499738 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.499824 kubelet[3245]: W0813 07:22:08.499757 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.499824 kubelet[3245]: E0813 07:22:08.499778 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.499993 kubelet[3245]: E0813 07:22:08.499979 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.500078 kubelet[3245]: W0813 07:22:08.499992 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.500078 kubelet[3245]: E0813 07:22:08.500013 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.500244 kubelet[3245]: E0813 07:22:08.500230 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.500244 kubelet[3245]: W0813 07:22:08.500243 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.500376 kubelet[3245]: E0813 07:22:08.500261 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.500507 kubelet[3245]: E0813 07:22:08.500493 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.500576 kubelet[3245]: W0813 07:22:08.500506 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.500576 kubelet[3245]: E0813 07:22:08.500526 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.500822 kubelet[3245]: E0813 07:22:08.500807 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.500875 kubelet[3245]: W0813 07:22:08.500822 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.500875 kubelet[3245]: E0813 07:22:08.500840 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.501049 kubelet[3245]: E0813 07:22:08.501037 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.501123 kubelet[3245]: W0813 07:22:08.501053 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.501123 kubelet[3245]: E0813 07:22:08.501087 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.501343 kubelet[3245]: E0813 07:22:08.501331 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.501392 kubelet[3245]: W0813 07:22:08.501347 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.501392 kubelet[3245]: E0813 07:22:08.501368 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.501649 kubelet[3245]: E0813 07:22:08.501634 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.501721 kubelet[3245]: W0813 07:22:08.501649 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.501721 kubelet[3245]: E0813 07:22:08.501670 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.501904 kubelet[3245]: E0813 07:22:08.501889 3245 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:22:08.501904 kubelet[3245]: W0813 07:22:08.501903 3245 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:22:08.502002 kubelet[3245]: E0813 07:22:08.501919 3245 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:22:08.619251 containerd[1920]: time="2025-08-13T07:22:08.619160409Z" level=info msg="StartContainer for \"162dd8242386a801ed86b3511cb02e64723c3dc63349420bb3fdd3e301ebd1b7\" returns successfully" Aug 13 07:22:08.718744 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-162dd8242386a801ed86b3511cb02e64723c3dc63349420bb3fdd3e301ebd1b7-rootfs.mount: Deactivated successfully. Aug 13 07:22:09.349917 kubelet[3245]: E0813 07:22:09.349773 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xnn8d" podUID="b90d45be-b93c-47a4-ad3d-74de560838fa" Aug 13 07:22:09.872152 containerd[1920]: time="2025-08-13T07:22:09.872113765Z" level=info msg="shim disconnected" id=162dd8242386a801ed86b3511cb02e64723c3dc63349420bb3fdd3e301ebd1b7 namespace=k8s.io Aug 13 07:22:09.872152 containerd[1920]: time="2025-08-13T07:22:09.872150301Z" level=warning msg="cleaning up after shim disconnected" id=162dd8242386a801ed86b3511cb02e64723c3dc63349420bb3fdd3e301ebd1b7 namespace=k8s.io Aug 13 07:22:09.872152 containerd[1920]: time="2025-08-13T07:22:09.872156868Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:22:10.415721 containerd[1920]: time="2025-08-13T07:22:10.415639427Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Aug 13 07:22:11.349660 kubelet[3245]: E0813 07:22:11.349505 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xnn8d" podUID="b90d45be-b93c-47a4-ad3d-74de560838fa" Aug 13 07:22:12.816947 containerd[1920]: time="2025-08-13T07:22:12.816896095Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:12.817179 containerd[1920]: time="2025-08-13T07:22:12.817063081Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.2: active requests=0, bytes read=70436221" Aug 13 07:22:12.817450 containerd[1920]: time="2025-08-13T07:22:12.817409721Z" level=info msg="ImageCreate event name:\"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:12.819027 containerd[1920]: time="2025-08-13T07:22:12.818587717Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:12.819458 containerd[1920]: time="2025-08-13T07:22:12.819444315Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.2\" with image id \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\", size \"71928924\" in 2.403724291s" Aug 13 07:22:12.819483 containerd[1920]: time="2025-08-13T07:22:12.819461125Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\"" Aug 13 07:22:12.820355 containerd[1920]: time="2025-08-13T07:22:12.820342164Z" level=info msg="CreateContainer within sandbox \"3410d01fc2075fca23a48d575638f24f9bed57f49e04f881b51dbe1f880902ae\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Aug 13 07:22:12.824998 containerd[1920]: time="2025-08-13T07:22:12.824956265Z" level=info msg="CreateContainer within sandbox \"3410d01fc2075fca23a48d575638f24f9bed57f49e04f881b51dbe1f880902ae\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"79502854e4b9cea0c15021fba0dc73ddbf5c153f7ceb37cee351400795388e6c\"" Aug 13 07:22:12.825243 containerd[1920]: time="2025-08-13T07:22:12.825196572Z" level=info msg="StartContainer for \"79502854e4b9cea0c15021fba0dc73ddbf5c153f7ceb37cee351400795388e6c\"" Aug 13 07:22:12.855215 containerd[1920]: time="2025-08-13T07:22:12.855192029Z" level=info msg="StartContainer for \"79502854e4b9cea0c15021fba0dc73ddbf5c153f7ceb37cee351400795388e6c\" returns successfully" Aug 13 07:22:13.349112 kubelet[3245]: E0813 07:22:13.349055 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-xnn8d" podUID="b90d45be-b93c-47a4-ad3d-74de560838fa" Aug 13 07:22:13.422330 containerd[1920]: time="2025-08-13T07:22:13.422304962Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 07:22:13.432734 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-79502854e4b9cea0c15021fba0dc73ddbf5c153f7ceb37cee351400795388e6c-rootfs.mount: Deactivated successfully. Aug 13 07:22:13.513870 kubelet[3245]: I0813 07:22:13.513771 3245 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Aug 13 07:22:13.737640 kubelet[3245]: I0813 07:22:13.737408 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/b2d55415-dd30-4e5a-9264-00bc2660bdba-calico-apiserver-certs\") pod \"calico-apiserver-59fbd88f7c-z2k5j\" (UID: \"b2d55415-dd30-4e5a-9264-00bc2660bdba\") " pod="calico-apiserver/calico-apiserver-59fbd88f7c-z2k5j" Aug 13 07:22:13.737640 kubelet[3245]: I0813 07:22:13.737521 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bknn7\" (UniqueName: \"kubernetes.io/projected/dd85a865-2c7f-4339-b9a7-90c5b37bd6d6-kube-api-access-bknn7\") pod \"goldmane-58fd7646b9-sjk28\" (UID: \"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6\") " pod="calico-system/goldmane-58fd7646b9-sjk28" Aug 13 07:22:13.737640 kubelet[3245]: I0813 07:22:13.737586 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/dd85a865-2c7f-4339-b9a7-90c5b37bd6d6-goldmane-key-pair\") pod \"goldmane-58fd7646b9-sjk28\" (UID: \"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6\") " pod="calico-system/goldmane-58fd7646b9-sjk28" Aug 13 07:22:13.737640 kubelet[3245]: I0813 07:22:13.737639 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/926b03e8-ac14-4763-b7d3-fa81c377f2dd-config-volume\") pod \"coredns-7c65d6cfc9-ggl67\" (UID: \"926b03e8-ac14-4763-b7d3-fa81c377f2dd\") " pod="kube-system/coredns-7c65d6cfc9-ggl67" Aug 13 07:22:13.738409 kubelet[3245]: I0813 07:22:13.737691 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-ca-bundle\") pod \"whisker-5f8549b86f-sqvnt\" (UID: \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\") " pod="calico-system/whisker-5f8549b86f-sqvnt" Aug 13 07:22:13.738409 kubelet[3245]: I0813 07:22:13.737743 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd85a865-2c7f-4339-b9a7-90c5b37bd6d6-config\") pod \"goldmane-58fd7646b9-sjk28\" (UID: \"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6\") " pod="calico-system/goldmane-58fd7646b9-sjk28" Aug 13 07:22:13.738409 kubelet[3245]: I0813 07:22:13.737813 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmq2w\" (UniqueName: \"kubernetes.io/projected/bb4a7045-3301-41f3-86a2-df653630104c-kube-api-access-cmq2w\") pod \"calico-apiserver-59fbd88f7c-qqszn\" (UID: \"bb4a7045-3301-41f3-86a2-df653630104c\") " pod="calico-apiserver/calico-apiserver-59fbd88f7c-qqszn" Aug 13 07:22:13.738409 kubelet[3245]: I0813 07:22:13.737892 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/89af5e87-c375-4eee-977b-cf953f1e3e5b-config-volume\") pod \"coredns-7c65d6cfc9-brr5j\" (UID: \"89af5e87-c375-4eee-977b-cf953f1e3e5b\") " pod="kube-system/coredns-7c65d6cfc9-brr5j" Aug 13 07:22:13.738409 kubelet[3245]: I0813 07:22:13.737947 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-backend-key-pair\") pod \"whisker-5f8549b86f-sqvnt\" (UID: \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\") " pod="calico-system/whisker-5f8549b86f-sqvnt" Aug 13 07:22:13.738953 kubelet[3245]: I0813 07:22:13.737998 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f80e5fda-a5d5-4f6a-9768-261c847bf3ee-tigera-ca-bundle\") pod \"calico-kube-controllers-694d7cc5f7-rl69w\" (UID: \"f80e5fda-a5d5-4f6a-9768-261c847bf3ee\") " pod="calico-system/calico-kube-controllers-694d7cc5f7-rl69w" Aug 13 07:22:13.738953 kubelet[3245]: I0813 07:22:13.738060 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/bb4a7045-3301-41f3-86a2-df653630104c-calico-apiserver-certs\") pod \"calico-apiserver-59fbd88f7c-qqszn\" (UID: \"bb4a7045-3301-41f3-86a2-df653630104c\") " pod="calico-apiserver/calico-apiserver-59fbd88f7c-qqszn" Aug 13 07:22:13.738953 kubelet[3245]: I0813 07:22:13.738137 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpz9v\" (UniqueName: \"kubernetes.io/projected/89af5e87-c375-4eee-977b-cf953f1e3e5b-kube-api-access-fpz9v\") pod \"coredns-7c65d6cfc9-brr5j\" (UID: \"89af5e87-c375-4eee-977b-cf953f1e3e5b\") " pod="kube-system/coredns-7c65d6cfc9-brr5j" Aug 13 07:22:13.738953 kubelet[3245]: I0813 07:22:13.738230 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rpcw\" (UniqueName: \"kubernetes.io/projected/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-kube-api-access-7rpcw\") pod \"whisker-5f8549b86f-sqvnt\" (UID: \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\") " pod="calico-system/whisker-5f8549b86f-sqvnt" Aug 13 07:22:13.738953 kubelet[3245]: I0813 07:22:13.738284 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd85a865-2c7f-4339-b9a7-90c5b37bd6d6-goldmane-ca-bundle\") pod \"goldmane-58fd7646b9-sjk28\" (UID: \"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6\") " pod="calico-system/goldmane-58fd7646b9-sjk28" Aug 13 07:22:13.739511 kubelet[3245]: I0813 07:22:13.738337 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62zf7\" (UniqueName: \"kubernetes.io/projected/926b03e8-ac14-4763-b7d3-fa81c377f2dd-kube-api-access-62zf7\") pod \"coredns-7c65d6cfc9-ggl67\" (UID: \"926b03e8-ac14-4763-b7d3-fa81c377f2dd\") " pod="kube-system/coredns-7c65d6cfc9-ggl67" Aug 13 07:22:13.739511 kubelet[3245]: I0813 07:22:13.738393 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjd8z\" (UniqueName: \"kubernetes.io/projected/f80e5fda-a5d5-4f6a-9768-261c847bf3ee-kube-api-access-zjd8z\") pod \"calico-kube-controllers-694d7cc5f7-rl69w\" (UID: \"f80e5fda-a5d5-4f6a-9768-261c847bf3ee\") " pod="calico-system/calico-kube-controllers-694d7cc5f7-rl69w" Aug 13 07:22:13.739511 kubelet[3245]: I0813 07:22:13.738444 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf7mx\" (UniqueName: \"kubernetes.io/projected/b2d55415-dd30-4e5a-9264-00bc2660bdba-kube-api-access-kf7mx\") pod \"calico-apiserver-59fbd88f7c-z2k5j\" (UID: \"b2d55415-dd30-4e5a-9264-00bc2660bdba\") " pod="calico-apiserver/calico-apiserver-59fbd88f7c-z2k5j" Aug 13 07:22:13.863913 containerd[1920]: time="2025-08-13T07:22:13.863879920Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-694d7cc5f7-rl69w,Uid:f80e5fda-a5d5-4f6a-9768-261c847bf3ee,Namespace:calico-system,Attempt:0,}" Aug 13 07:22:13.865326 containerd[1920]: time="2025-08-13T07:22:13.863950144Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-brr5j,Uid:89af5e87-c375-4eee-977b-cf953f1e3e5b,Namespace:kube-system,Attempt:0,}" Aug 13 07:22:13.865721 containerd[1920]: time="2025-08-13T07:22:13.865698736Z" level=info msg="shim disconnected" id=79502854e4b9cea0c15021fba0dc73ddbf5c153f7ceb37cee351400795388e6c namespace=k8s.io Aug 13 07:22:13.865766 containerd[1920]: time="2025-08-13T07:22:13.865721486Z" level=warning msg="cleaning up after shim disconnected" id=79502854e4b9cea0c15021fba0dc73ddbf5c153f7ceb37cee351400795388e6c namespace=k8s.io Aug 13 07:22:13.865766 containerd[1920]: time="2025-08-13T07:22:13.865729567Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:22:13.866312 containerd[1920]: time="2025-08-13T07:22:13.866297398Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-ggl67,Uid:926b03e8-ac14-4763-b7d3-fa81c377f2dd,Namespace:kube-system,Attempt:0,}" Aug 13 07:22:13.869785 containerd[1920]: time="2025-08-13T07:22:13.869765103Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-z2k5j,Uid:b2d55415-dd30-4e5a-9264-00bc2660bdba,Namespace:calico-apiserver,Attempt:0,}" Aug 13 07:22:13.871204 containerd[1920]: time="2025-08-13T07:22:13.871185728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-qqszn,Uid:bb4a7045-3301-41f3-86a2-df653630104c,Namespace:calico-apiserver,Attempt:0,}" Aug 13 07:22:13.872532 containerd[1920]: time="2025-08-13T07:22:13.872517132Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-sjk28,Uid:dd85a865-2c7f-4339-b9a7-90c5b37bd6d6,Namespace:calico-system,Attempt:0,}" Aug 13 07:22:13.874951 containerd[1920]: time="2025-08-13T07:22:13.874934380Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5f8549b86f-sqvnt,Uid:e8f971dd-b3c2-4184-b9bd-768a0ccdaa22,Namespace:calico-system,Attempt:0,}" Aug 13 07:22:13.903961 containerd[1920]: time="2025-08-13T07:22:13.903929032Z" level=error msg="Failed to destroy network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904209 containerd[1920]: time="2025-08-13T07:22:13.904195621Z" level=error msg="encountered an error cleaning up failed sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904239 containerd[1920]: time="2025-08-13T07:22:13.904226002Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-brr5j,Uid:89af5e87-c375-4eee-977b-cf953f1e3e5b,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904404 kubelet[3245]: E0813 07:22:13.904381 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904451 kubelet[3245]: E0813 07:22:13.904431 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-brr5j" Aug 13 07:22:13.904451 kubelet[3245]: E0813 07:22:13.904444 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-brr5j" Aug 13 07:22:13.904494 kubelet[3245]: E0813 07:22:13.904473 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-brr5j_kube-system(89af5e87-c375-4eee-977b-cf953f1e3e5b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-brr5j_kube-system(89af5e87-c375-4eee-977b-cf953f1e3e5b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-brr5j" podUID="89af5e87-c375-4eee-977b-cf953f1e3e5b" Aug 13 07:22:13.904591 containerd[1920]: time="2025-08-13T07:22:13.904567487Z" level=error msg="Failed to destroy network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904617 containerd[1920]: time="2025-08-13T07:22:13.904592568Z" level=error msg="Failed to destroy network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904805 containerd[1920]: time="2025-08-13T07:22:13.904787713Z" level=error msg="encountered an error cleaning up failed sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904838 containerd[1920]: time="2025-08-13T07:22:13.904820771Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-694d7cc5f7-rl69w,Uid:f80e5fda-a5d5-4f6a-9768-261c847bf3ee,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904875 containerd[1920]: time="2025-08-13T07:22:13.904830597Z" level=error msg="encountered an error cleaning up failed sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904902 containerd[1920]: time="2025-08-13T07:22:13.904885869Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-ggl67,Uid:926b03e8-ac14-4763-b7d3-fa81c377f2dd,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904958 kubelet[3245]: E0813 07:22:13.904945 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904983 kubelet[3245]: E0813 07:22:13.904956 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.904983 kubelet[3245]: E0813 07:22:13.904966 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-694d7cc5f7-rl69w" Aug 13 07:22:13.904983 kubelet[3245]: E0813 07:22:13.904977 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-694d7cc5f7-rl69w" Aug 13 07:22:13.905043 kubelet[3245]: E0813 07:22:13.904976 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-ggl67" Aug 13 07:22:13.905043 kubelet[3245]: E0813 07:22:13.904996 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-694d7cc5f7-rl69w_calico-system(f80e5fda-a5d5-4f6a-9768-261c847bf3ee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-694d7cc5f7-rl69w_calico-system(f80e5fda-a5d5-4f6a-9768-261c847bf3ee)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-694d7cc5f7-rl69w" podUID="f80e5fda-a5d5-4f6a-9768-261c847bf3ee" Aug 13 07:22:13.905043 kubelet[3245]: E0813 07:22:13.905010 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-ggl67" Aug 13 07:22:13.905127 kubelet[3245]: E0813 07:22:13.905037 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-ggl67_kube-system(926b03e8-ac14-4763-b7d3-fa81c377f2dd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-ggl67_kube-system(926b03e8-ac14-4763-b7d3-fa81c377f2dd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-ggl67" podUID="926b03e8-ac14-4763-b7d3-fa81c377f2dd" Aug 13 07:22:13.912317 containerd[1920]: time="2025-08-13T07:22:13.912252180Z" level=error msg="Failed to destroy network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912442 containerd[1920]: time="2025-08-13T07:22:13.912421878Z" level=error msg="Failed to destroy network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912472 containerd[1920]: time="2025-08-13T07:22:13.912446496Z" level=error msg="encountered an error cleaning up failed sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912493 containerd[1920]: time="2025-08-13T07:22:13.912474039Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-sjk28,Uid:dd85a865-2c7f-4339-b9a7-90c5b37bd6d6,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912583 containerd[1920]: time="2025-08-13T07:22:13.912570863Z" level=error msg="encountered an error cleaning up failed sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912608 containerd[1920]: time="2025-08-13T07:22:13.912592168Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-qqszn,Uid:bb4a7045-3301-41f3-86a2-df653630104c,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912639 kubelet[3245]: E0813 07:22:13.912582 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912639 kubelet[3245]: E0813 07:22:13.912617 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-sjk28" Aug 13 07:22:13.912639 kubelet[3245]: E0813 07:22:13.912631 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-sjk28" Aug 13 07:22:13.912704 kubelet[3245]: E0813 07:22:13.912656 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.912704 kubelet[3245]: E0813 07:22:13.912661 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-58fd7646b9-sjk28_calico-system(dd85a865-2c7f-4339-b9a7-90c5b37bd6d6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-58fd7646b9-sjk28_calico-system(dd85a865-2c7f-4339-b9a7-90c5b37bd6d6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-sjk28" podUID="dd85a865-2c7f-4339-b9a7-90c5b37bd6d6" Aug 13 07:22:13.912704 kubelet[3245]: E0813 07:22:13.912677 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59fbd88f7c-qqszn" Aug 13 07:22:13.912773 kubelet[3245]: E0813 07:22:13.912693 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59fbd88f7c-qqszn" Aug 13 07:22:13.912773 kubelet[3245]: E0813 07:22:13.912714 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-59fbd88f7c-qqszn_calico-apiserver(bb4a7045-3301-41f3-86a2-df653630104c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-59fbd88f7c-qqszn_calico-apiserver(bb4a7045-3301-41f3-86a2-df653630104c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59fbd88f7c-qqszn" podUID="bb4a7045-3301-41f3-86a2-df653630104c" Aug 13 07:22:13.914358 containerd[1920]: time="2025-08-13T07:22:13.914336409Z" level=error msg="Failed to destroy network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914500 containerd[1920]: time="2025-08-13T07:22:13.914487851Z" level=error msg="encountered an error cleaning up failed sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914531 containerd[1920]: time="2025-08-13T07:22:13.914515867Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-z2k5j,Uid:b2d55415-dd30-4e5a-9264-00bc2660bdba,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914575 containerd[1920]: time="2025-08-13T07:22:13.914556867Z" level=error msg="Failed to destroy network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914600 kubelet[3245]: E0813 07:22:13.914585 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914627 kubelet[3245]: E0813 07:22:13.914605 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59fbd88f7c-z2k5j" Aug 13 07:22:13.914627 kubelet[3245]: E0813 07:22:13.914616 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-59fbd88f7c-z2k5j" Aug 13 07:22:13.914668 kubelet[3245]: E0813 07:22:13.914635 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-59fbd88f7c-z2k5j_calico-apiserver(b2d55415-dd30-4e5a-9264-00bc2660bdba)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-59fbd88f7c-z2k5j_calico-apiserver(b2d55415-dd30-4e5a-9264-00bc2660bdba)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59fbd88f7c-z2k5j" podUID="b2d55415-dd30-4e5a-9264-00bc2660bdba" Aug 13 07:22:13.914711 containerd[1920]: time="2025-08-13T07:22:13.914698345Z" level=error msg="encountered an error cleaning up failed sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914732 containerd[1920]: time="2025-08-13T07:22:13.914718582Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5f8549b86f-sqvnt,Uid:e8f971dd-b3c2-4184-b9bd-768a0ccdaa22,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914790 kubelet[3245]: E0813 07:22:13.914778 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:13.914817 kubelet[3245]: E0813 07:22:13.914801 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5f8549b86f-sqvnt" Aug 13 07:22:13.914817 kubelet[3245]: E0813 07:22:13.914811 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5f8549b86f-sqvnt" Aug 13 07:22:13.914855 kubelet[3245]: E0813 07:22:13.914831 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-5f8549b86f-sqvnt_calico-system(e8f971dd-b3c2-4184-b9bd-768a0ccdaa22)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-5f8549b86f-sqvnt_calico-system(e8f971dd-b3c2-4184-b9bd-768a0ccdaa22)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5f8549b86f-sqvnt" podUID="e8f971dd-b3c2-4184-b9bd-768a0ccdaa22" Aug 13 07:22:14.424610 kubelet[3245]: I0813 07:22:14.424513 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:14.426038 containerd[1920]: time="2025-08-13T07:22:14.425965730Z" level=info msg="StopPodSandbox for \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\"" Aug 13 07:22:14.426474 containerd[1920]: time="2025-08-13T07:22:14.426422815Z" level=info msg="Ensure that sandbox 1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c in task-service has been cleanup successfully" Aug 13 07:22:14.430591 kubelet[3245]: I0813 07:22:14.430516 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:14.430885 containerd[1920]: time="2025-08-13T07:22:14.430809210Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Aug 13 07:22:14.431907 containerd[1920]: time="2025-08-13T07:22:14.431827096Z" level=info msg="StopPodSandbox for \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\"" Aug 13 07:22:14.432314 containerd[1920]: time="2025-08-13T07:22:14.432301281Z" level=info msg="Ensure that sandbox f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a in task-service has been cleanup successfully" Aug 13 07:22:14.432368 kubelet[3245]: I0813 07:22:14.432358 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:14.432620 containerd[1920]: time="2025-08-13T07:22:14.432602725Z" level=info msg="StopPodSandbox for \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\"" Aug 13 07:22:14.432734 containerd[1920]: time="2025-08-13T07:22:14.432720944Z" level=info msg="Ensure that sandbox 47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983 in task-service has been cleanup successfully" Aug 13 07:22:14.432795 kubelet[3245]: I0813 07:22:14.432786 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:14.433060 containerd[1920]: time="2025-08-13T07:22:14.433048754Z" level=info msg="StopPodSandbox for \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\"" Aug 13 07:22:14.433169 containerd[1920]: time="2025-08-13T07:22:14.433155757Z" level=info msg="Ensure that sandbox c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1 in task-service has been cleanup successfully" Aug 13 07:22:14.433281 kubelet[3245]: I0813 07:22:14.433273 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:14.433547 containerd[1920]: time="2025-08-13T07:22:14.433529029Z" level=info msg="StopPodSandbox for \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\"" Aug 13 07:22:14.433672 containerd[1920]: time="2025-08-13T07:22:14.433654363Z" level=info msg="Ensure that sandbox 47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56 in task-service has been cleanup successfully" Aug 13 07:22:14.433895 kubelet[3245]: I0813 07:22:14.433879 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:14.434357 containerd[1920]: time="2025-08-13T07:22:14.434326557Z" level=info msg="StopPodSandbox for \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\"" Aug 13 07:22:14.434582 kubelet[3245]: I0813 07:22:14.434481 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:14.434666 containerd[1920]: time="2025-08-13T07:22:14.434507817Z" level=info msg="Ensure that sandbox a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451 in task-service has been cleanup successfully" Aug 13 07:22:14.434880 containerd[1920]: time="2025-08-13T07:22:14.434866103Z" level=info msg="StopPodSandbox for \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\"" Aug 13 07:22:14.434979 containerd[1920]: time="2025-08-13T07:22:14.434965815Z" level=info msg="Ensure that sandbox 74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296 in task-service has been cleanup successfully" Aug 13 07:22:14.447744 containerd[1920]: time="2025-08-13T07:22:14.447714003Z" level=error msg="StopPodSandbox for \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\" failed" error="failed to destroy network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:14.447853 kubelet[3245]: E0813 07:22:14.447835 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:14.447901 kubelet[3245]: E0813 07:22:14.447868 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c"} Aug 13 07:22:14.447928 kubelet[3245]: E0813 07:22:14.447914 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f80e5fda-a5d5-4f6a-9768-261c847bf3ee\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:14.447972 kubelet[3245]: E0813 07:22:14.447929 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f80e5fda-a5d5-4f6a-9768-261c847bf3ee\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-694d7cc5f7-rl69w" podUID="f80e5fda-a5d5-4f6a-9768-261c847bf3ee" Aug 13 07:22:14.448017 containerd[1920]: time="2025-08-13T07:22:14.447927073Z" level=error msg="StopPodSandbox for \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\" failed" error="failed to destroy network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:14.448043 kubelet[3245]: E0813 07:22:14.448000 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:14.448043 kubelet[3245]: E0813 07:22:14.448017 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a"} Aug 13 07:22:14.448043 kubelet[3245]: E0813 07:22:14.448030 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:14.448043 kubelet[3245]: E0813 07:22:14.448040 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-sjk28" podUID="dd85a865-2c7f-4339-b9a7-90c5b37bd6d6" Aug 13 07:22:14.448356 containerd[1920]: time="2025-08-13T07:22:14.448341600Z" level=error msg="StopPodSandbox for \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\" failed" error="failed to destroy network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:14.448423 kubelet[3245]: E0813 07:22:14.448406 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:14.448464 kubelet[3245]: E0813 07:22:14.448429 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983"} Aug 13 07:22:14.448464 kubelet[3245]: E0813 07:22:14.448453 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"b2d55415-dd30-4e5a-9264-00bc2660bdba\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:14.448526 kubelet[3245]: E0813 07:22:14.448472 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"b2d55415-dd30-4e5a-9264-00bc2660bdba\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59fbd88f7c-z2k5j" podUID="b2d55415-dd30-4e5a-9264-00bc2660bdba" Aug 13 07:22:14.449674 containerd[1920]: time="2025-08-13T07:22:14.449608081Z" level=error msg="StopPodSandbox for \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\" failed" error="failed to destroy network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:14.449717 kubelet[3245]: E0813 07:22:14.449705 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:14.449746 kubelet[3245]: E0813 07:22:14.449721 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451"} Aug 13 07:22:14.449746 kubelet[3245]: E0813 07:22:14.449739 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"bb4a7045-3301-41f3-86a2-df653630104c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:14.449795 kubelet[3245]: E0813 07:22:14.449749 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"bb4a7045-3301-41f3-86a2-df653630104c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-59fbd88f7c-qqszn" podUID="bb4a7045-3301-41f3-86a2-df653630104c" Aug 13 07:22:14.449940 containerd[1920]: time="2025-08-13T07:22:14.449925320Z" level=error msg="StopPodSandbox for \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\" failed" error="failed to destroy network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:14.449998 kubelet[3245]: E0813 07:22:14.449989 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:14.450025 kubelet[3245]: E0813 07:22:14.449999 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296"} Aug 13 07:22:14.450025 kubelet[3245]: E0813 07:22:14.450012 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"89af5e87-c375-4eee-977b-cf953f1e3e5b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:14.450025 kubelet[3245]: E0813 07:22:14.450021 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"89af5e87-c375-4eee-977b-cf953f1e3e5b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-brr5j" podUID="89af5e87-c375-4eee-977b-cf953f1e3e5b" Aug 13 07:22:14.450108 containerd[1920]: time="2025-08-13T07:22:14.450090825Z" level=error msg="StopPodSandbox for \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\" failed" error="failed to destroy network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:14.450129 containerd[1920]: time="2025-08-13T07:22:14.450117138Z" level=error msg="StopPodSandbox for \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\" failed" error="failed to destroy network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:14.450186 kubelet[3245]: E0813 07:22:14.450173 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:14.450226 kubelet[3245]: E0813 07:22:14.450190 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56"} Aug 13 07:22:14.450226 kubelet[3245]: E0813 07:22:14.450205 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:14.450226 kubelet[3245]: E0813 07:22:14.450212 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:14.450226 kubelet[3245]: E0813 07:22:14.450221 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1"} Aug 13 07:22:14.450345 kubelet[3245]: E0813 07:22:14.450231 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5f8549b86f-sqvnt" podUID="e8f971dd-b3c2-4184-b9bd-768a0ccdaa22" Aug 13 07:22:14.450345 kubelet[3245]: E0813 07:22:14.450236 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"926b03e8-ac14-4763-b7d3-fa81c377f2dd\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:14.450345 kubelet[3245]: E0813 07:22:14.450256 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"926b03e8-ac14-4763-b7d3-fa81c377f2dd\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-ggl67" podUID="926b03e8-ac14-4763-b7d3-fa81c377f2dd" Aug 13 07:22:14.851192 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1-shm.mount: Deactivated successfully. Aug 13 07:22:14.851263 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296-shm.mount: Deactivated successfully. Aug 13 07:22:14.851315 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c-shm.mount: Deactivated successfully. Aug 13 07:22:15.350930 containerd[1920]: time="2025-08-13T07:22:15.350904992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xnn8d,Uid:b90d45be-b93c-47a4-ad3d-74de560838fa,Namespace:calico-system,Attempt:0,}" Aug 13 07:22:15.378168 containerd[1920]: time="2025-08-13T07:22:15.378140198Z" level=error msg="Failed to destroy network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:15.378426 containerd[1920]: time="2025-08-13T07:22:15.378386195Z" level=error msg="encountered an error cleaning up failed sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:15.378426 containerd[1920]: time="2025-08-13T07:22:15.378413798Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xnn8d,Uid:b90d45be-b93c-47a4-ad3d-74de560838fa,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:15.378610 kubelet[3245]: E0813 07:22:15.378562 3245 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:15.378643 kubelet[3245]: E0813 07:22:15.378605 3245 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-xnn8d" Aug 13 07:22:15.378643 kubelet[3245]: E0813 07:22:15.378620 3245 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-xnn8d" Aug 13 07:22:15.378685 kubelet[3245]: E0813 07:22:15.378648 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-xnn8d_calico-system(b90d45be-b93c-47a4-ad3d-74de560838fa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-xnn8d_calico-system(b90d45be-b93c-47a4-ad3d-74de560838fa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-xnn8d" podUID="b90d45be-b93c-47a4-ad3d-74de560838fa" Aug 13 07:22:15.379811 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e-shm.mount: Deactivated successfully. Aug 13 07:22:15.438610 kubelet[3245]: I0813 07:22:15.438524 3245 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:15.439663 containerd[1920]: time="2025-08-13T07:22:15.439565851Z" level=info msg="StopPodSandbox for \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\"" Aug 13 07:22:15.440031 containerd[1920]: time="2025-08-13T07:22:15.439964468Z" level=info msg="Ensure that sandbox 7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e in task-service has been cleanup successfully" Aug 13 07:22:15.495824 containerd[1920]: time="2025-08-13T07:22:15.495740490Z" level=error msg="StopPodSandbox for \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\" failed" error="failed to destroy network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:22:15.496040 kubelet[3245]: E0813 07:22:15.495989 3245 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:15.496133 kubelet[3245]: E0813 07:22:15.496059 3245 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e"} Aug 13 07:22:15.496199 kubelet[3245]: E0813 07:22:15.496129 3245 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"b90d45be-b93c-47a4-ad3d-74de560838fa\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:22:15.496199 kubelet[3245]: E0813 07:22:15.496166 3245 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"b90d45be-b93c-47a4-ad3d-74de560838fa\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-xnn8d" podUID="b90d45be-b93c-47a4-ad3d-74de560838fa" Aug 13 07:22:17.889998 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2388743690.mount: Deactivated successfully. Aug 13 07:22:17.909803 containerd[1920]: time="2025-08-13T07:22:17.909756032Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:17.909977 containerd[1920]: time="2025-08-13T07:22:17.909949332Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.2: active requests=0, bytes read=158500163" Aug 13 07:22:17.910360 containerd[1920]: time="2025-08-13T07:22:17.910322738Z" level=info msg="ImageCreate event name:\"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:17.911174 containerd[1920]: time="2025-08-13T07:22:17.911131831Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:17.911544 containerd[1920]: time="2025-08-13T07:22:17.911508038Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.2\" with image id \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\", size \"158500025\" in 3.480596148s" Aug 13 07:22:17.911544 containerd[1920]: time="2025-08-13T07:22:17.911523971Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\"" Aug 13 07:22:17.914858 containerd[1920]: time="2025-08-13T07:22:17.914843053Z" level=info msg="CreateContainer within sandbox \"3410d01fc2075fca23a48d575638f24f9bed57f49e04f881b51dbe1f880902ae\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Aug 13 07:22:17.920838 containerd[1920]: time="2025-08-13T07:22:17.920792832Z" level=info msg="CreateContainer within sandbox \"3410d01fc2075fca23a48d575638f24f9bed57f49e04f881b51dbe1f880902ae\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"1cedbed0840165d16146be6c0c822929af859113bf008413d36ca5b247703336\"" Aug 13 07:22:17.921094 containerd[1920]: time="2025-08-13T07:22:17.921077482Z" level=info msg="StartContainer for \"1cedbed0840165d16146be6c0c822929af859113bf008413d36ca5b247703336\"" Aug 13 07:22:17.966477 containerd[1920]: time="2025-08-13T07:22:17.966422688Z" level=info msg="StartContainer for \"1cedbed0840165d16146be6c0c822929af859113bf008413d36ca5b247703336\" returns successfully" Aug 13 07:22:18.036430 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Aug 13 07:22:18.036491 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Aug 13 07:22:18.073827 containerd[1920]: time="2025-08-13T07:22:18.073799570Z" level=info msg="StopPodSandbox for \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\"" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.098 [INFO][4852] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.098 [INFO][4852] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" iface="eth0" netns="/var/run/netns/cni-57c9d35e-06a2-9df6-5db8-fbdcac44d055" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.098 [INFO][4852] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" iface="eth0" netns="/var/run/netns/cni-57c9d35e-06a2-9df6-5db8-fbdcac44d055" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.098 [INFO][4852] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" iface="eth0" netns="/var/run/netns/cni-57c9d35e-06a2-9df6-5db8-fbdcac44d055" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.098 [INFO][4852] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.098 [INFO][4852] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.109 [INFO][4881] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.109 [INFO][4881] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.109 [INFO][4881] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.112 [WARNING][4881] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.112 [INFO][4881] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.113 [INFO][4881] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:18.115914 containerd[1920]: 2025-08-13 07:22:18.115 [INFO][4852] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:18.116206 containerd[1920]: time="2025-08-13T07:22:18.115987265Z" level=info msg="TearDown network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\" successfully" Aug 13 07:22:18.116206 containerd[1920]: time="2025-08-13T07:22:18.116004013Z" level=info msg="StopPodSandbox for \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\" returns successfully" Aug 13 07:22:18.265830 kubelet[3245]: I0813 07:22:18.265582 3245 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-ca-bundle\") pod \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\" (UID: \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\") " Aug 13 07:22:18.265830 kubelet[3245]: I0813 07:22:18.265689 3245 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-backend-key-pair\") pod \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\" (UID: \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\") " Aug 13 07:22:18.266819 kubelet[3245]: I0813 07:22:18.265817 3245 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rpcw\" (UniqueName: \"kubernetes.io/projected/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-kube-api-access-7rpcw\") pod \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\" (UID: \"e8f971dd-b3c2-4184-b9bd-768a0ccdaa22\") " Aug 13 07:22:18.266819 kubelet[3245]: I0813 07:22:18.266691 3245 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "e8f971dd-b3c2-4184-b9bd-768a0ccdaa22" (UID: "e8f971dd-b3c2-4184-b9bd-768a0ccdaa22"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Aug 13 07:22:18.271964 kubelet[3245]: I0813 07:22:18.271865 3245 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-kube-api-access-7rpcw" (OuterVolumeSpecName: "kube-api-access-7rpcw") pod "e8f971dd-b3c2-4184-b9bd-768a0ccdaa22" (UID: "e8f971dd-b3c2-4184-b9bd-768a0ccdaa22"). InnerVolumeSpecName "kube-api-access-7rpcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 07:22:18.272168 kubelet[3245]: I0813 07:22:18.271955 3245 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "e8f971dd-b3c2-4184-b9bd-768a0ccdaa22" (UID: "e8f971dd-b3c2-4184-b9bd-768a0ccdaa22"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Aug 13 07:22:18.366516 kubelet[3245]: I0813 07:22:18.366426 3245 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rpcw\" (UniqueName: \"kubernetes.io/projected/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-kube-api-access-7rpcw\") on node \"ci-4081.3.5-a-3ca3f1276f\" DevicePath \"\"" Aug 13 07:22:18.366516 kubelet[3245]: I0813 07:22:18.366488 3245 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-ca-bundle\") on node \"ci-4081.3.5-a-3ca3f1276f\" DevicePath \"\"" Aug 13 07:22:18.366516 kubelet[3245]: I0813 07:22:18.366518 3245 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22-whisker-backend-key-pair\") on node \"ci-4081.3.5-a-3ca3f1276f\" DevicePath \"\"" Aug 13 07:22:18.488366 kubelet[3245]: I0813 07:22:18.488255 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-s8brb" podStartSLOduration=1.74190283 podStartE2EDuration="14.488219525s" podCreationTimestamp="2025-08-13 07:22:04 +0000 UTC" firstStartedPulling="2025-08-13 07:22:05.165580901 +0000 UTC m=+16.888542785" lastFinishedPulling="2025-08-13 07:22:17.911897594 +0000 UTC m=+29.634859480" observedRunningTime="2025-08-13 07:22:18.487660806 +0000 UTC m=+30.210622753" watchObservedRunningTime="2025-08-13 07:22:18.488219525 +0000 UTC m=+30.211181453" Aug 13 07:22:18.668838 kubelet[3245]: I0813 07:22:18.668718 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e45875f3-a021-46e7-9ae7-d616885098af-whisker-backend-key-pair\") pod \"whisker-8dbb44b87-hjlnr\" (UID: \"e45875f3-a021-46e7-9ae7-d616885098af\") " pod="calico-system/whisker-8dbb44b87-hjlnr" Aug 13 07:22:18.668838 kubelet[3245]: I0813 07:22:18.668843 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k7sf\" (UniqueName: \"kubernetes.io/projected/e45875f3-a021-46e7-9ae7-d616885098af-kube-api-access-4k7sf\") pod \"whisker-8dbb44b87-hjlnr\" (UID: \"e45875f3-a021-46e7-9ae7-d616885098af\") " pod="calico-system/whisker-8dbb44b87-hjlnr" Aug 13 07:22:18.669219 kubelet[3245]: I0813 07:22:18.668939 3245 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e45875f3-a021-46e7-9ae7-d616885098af-whisker-ca-bundle\") pod \"whisker-8dbb44b87-hjlnr\" (UID: \"e45875f3-a021-46e7-9ae7-d616885098af\") " pod="calico-system/whisker-8dbb44b87-hjlnr" Aug 13 07:22:18.843143 containerd[1920]: time="2025-08-13T07:22:18.843011559Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-8dbb44b87-hjlnr,Uid:e45875f3-a021-46e7-9ae7-d616885098af,Namespace:calico-system,Attempt:0,}" Aug 13 07:22:18.892644 systemd[1]: run-netns-cni\x2d57c9d35e\x2d06a2\x2d9df6\x2d5db8\x2dfbdcac44d055.mount: Deactivated successfully. Aug 13 07:22:18.892735 systemd[1]: var-lib-kubelet-pods-e8f971dd\x2db3c2\x2d4184\x2db9bd\x2d768a0ccdaa22-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d7rpcw.mount: Deactivated successfully. Aug 13 07:22:18.892802 systemd[1]: var-lib-kubelet-pods-e8f971dd\x2db3c2\x2d4184\x2db9bd\x2d768a0ccdaa22-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Aug 13 07:22:18.899356 systemd-networkd[1561]: calieb83cafd72b: Link UP Aug 13 07:22:18.899473 systemd-networkd[1561]: calieb83cafd72b: Gained carrier Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.858 [INFO][4911] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.865 [INFO][4911] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0 whisker-8dbb44b87- calico-system e45875f3-a021-46e7-9ae7-d616885098af 856 0 2025-08-13 07:22:18 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:8dbb44b87 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f whisker-8dbb44b87-hjlnr eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calieb83cafd72b [] [] }} ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.865 [INFO][4911] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.877 [INFO][4934] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" HandleID="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.877 [INFO][4934] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" HandleID="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f750), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"whisker-8dbb44b87-hjlnr", "timestamp":"2025-08-13 07:22:18.877334043 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.877 [INFO][4934] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.877 [INFO][4934] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.877 [INFO][4934] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.881 [INFO][4934] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.884 [INFO][4934] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.886 [INFO][4934] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.887 [INFO][4934] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.888 [INFO][4934] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.888 [INFO][4934] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.889 [INFO][4934] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96 Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.891 [INFO][4934] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.894 [INFO][4934] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.65/26] block=192.168.121.64/26 handle="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.894 [INFO][4934] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.65/26] handle="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.894 [INFO][4934] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:18.905593 containerd[1920]: 2025-08-13 07:22:18.894 [INFO][4934] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.65/26] IPv6=[] ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" HandleID="k8s-pod-network.78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" Aug 13 07:22:18.906096 containerd[1920]: 2025-08-13 07:22:18.895 [INFO][4911] cni-plugin/k8s.go 418: Populated endpoint ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0", GenerateName:"whisker-8dbb44b87-", Namespace:"calico-system", SelfLink:"", UID:"e45875f3-a021-46e7-9ae7-d616885098af", ResourceVersion:"856", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"8dbb44b87", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"whisker-8dbb44b87-hjlnr", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.121.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calieb83cafd72b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:18.906096 containerd[1920]: 2025-08-13 07:22:18.895 [INFO][4911] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.65/32] ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" Aug 13 07:22:18.906096 containerd[1920]: 2025-08-13 07:22:18.895 [INFO][4911] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieb83cafd72b ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" Aug 13 07:22:18.906096 containerd[1920]: 2025-08-13 07:22:18.899 [INFO][4911] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" Aug 13 07:22:18.906096 containerd[1920]: 2025-08-13 07:22:18.900 [INFO][4911] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0", GenerateName:"whisker-8dbb44b87-", Namespace:"calico-system", SelfLink:"", UID:"e45875f3-a021-46e7-9ae7-d616885098af", ResourceVersion:"856", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"8dbb44b87", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96", Pod:"whisker-8dbb44b87-hjlnr", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.121.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calieb83cafd72b", MAC:"e2:a3:39:7b:ff:df", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:18.906096 containerd[1920]: 2025-08-13 07:22:18.904 [INFO][4911] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96" Namespace="calico-system" Pod="whisker-8dbb44b87-hjlnr" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--8dbb44b87--hjlnr-eth0" Aug 13 07:22:18.914090 containerd[1920]: time="2025-08-13T07:22:18.914002465Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:18.914090 containerd[1920]: time="2025-08-13T07:22:18.914029942Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:18.914090 containerd[1920]: time="2025-08-13T07:22:18.914037024Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:18.914338 containerd[1920]: time="2025-08-13T07:22:18.914085409Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:18.958638 containerd[1920]: time="2025-08-13T07:22:18.958613874Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-8dbb44b87-hjlnr,Uid:e45875f3-a021-46e7-9ae7-d616885098af,Namespace:calico-system,Attempt:0,} returns sandbox id \"78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96\"" Aug 13 07:22:18.959380 containerd[1920]: time="2025-08-13T07:22:18.959367603Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Aug 13 07:22:19.459986 kubelet[3245]: I0813 07:22:19.459928 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:20.266316 systemd-networkd[1561]: calieb83cafd72b: Gained IPv6LL Aug 13 07:22:20.354164 kubelet[3245]: I0813 07:22:20.354097 3245 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f971dd-b3c2-4184-b9bd-768a0ccdaa22" path="/var/lib/kubelet/pods/e8f971dd-b3c2-4184-b9bd-768a0ccdaa22/volumes" Aug 13 07:22:20.468099 containerd[1920]: time="2025-08-13T07:22:20.468027952Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:20.468339 containerd[1920]: time="2025-08-13T07:22:20.468300959Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.2: active requests=0, bytes read=4661207" Aug 13 07:22:20.468648 containerd[1920]: time="2025-08-13T07:22:20.468610023Z" level=info msg="ImageCreate event name:\"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:20.469688 containerd[1920]: time="2025-08-13T07:22:20.469648095Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:20.470172 containerd[1920]: time="2025-08-13T07:22:20.470114901Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.2\" with image id \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\", size \"6153902\" in 1.510728386s" Aug 13 07:22:20.470172 containerd[1920]: time="2025-08-13T07:22:20.470129989Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\"" Aug 13 07:22:20.471824 containerd[1920]: time="2025-08-13T07:22:20.471810573Z" level=info msg="CreateContainer within sandbox \"78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Aug 13 07:22:20.494353 containerd[1920]: time="2025-08-13T07:22:20.494306729Z" level=info msg="CreateContainer within sandbox \"78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"8a8d6309f99e45dfbae3cf27c9b310eba4c864ec5549afac3826ef13e5f051b9\"" Aug 13 07:22:20.494715 containerd[1920]: time="2025-08-13T07:22:20.494700912Z" level=info msg="StartContainer for \"8a8d6309f99e45dfbae3cf27c9b310eba4c864ec5549afac3826ef13e5f051b9\"" Aug 13 07:22:20.541260 containerd[1920]: time="2025-08-13T07:22:20.541238017Z" level=info msg="StartContainer for \"8a8d6309f99e45dfbae3cf27c9b310eba4c864ec5549afac3826ef13e5f051b9\" returns successfully" Aug 13 07:22:20.541809 containerd[1920]: time="2025-08-13T07:22:20.541795002Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Aug 13 07:22:22.504345 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2333269742.mount: Deactivated successfully. Aug 13 07:22:22.508359 containerd[1920]: time="2025-08-13T07:22:22.508315733Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:22.508530 containerd[1920]: time="2025-08-13T07:22:22.508462911Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.2: active requests=0, bytes read=33083477" Aug 13 07:22:22.508858 containerd[1920]: time="2025-08-13T07:22:22.508798305Z" level=info msg="ImageCreate event name:\"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:22.509883 containerd[1920]: time="2025-08-13T07:22:22.509871400Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:22.510332 containerd[1920]: time="2025-08-13T07:22:22.510315245Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" with image id \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\", size \"33083307\" in 1.968500902s" Aug 13 07:22:22.510368 containerd[1920]: time="2025-08-13T07:22:22.510331815Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\"" Aug 13 07:22:22.511256 containerd[1920]: time="2025-08-13T07:22:22.511236440Z" level=info msg="CreateContainer within sandbox \"78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Aug 13 07:22:22.515133 containerd[1920]: time="2025-08-13T07:22:22.515115958Z" level=info msg="CreateContainer within sandbox \"78a8d828090cbfbbda3f881e0eae02c0d12e307ad7eddbc39646640db27c8b96\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"4d92366739885ea706be665b5820656471dedf5a6c81ff21acb017d952ebafff\"" Aug 13 07:22:22.515370 containerd[1920]: time="2025-08-13T07:22:22.515355937Z" level=info msg="StartContainer for \"4d92366739885ea706be665b5820656471dedf5a6c81ff21acb017d952ebafff\"" Aug 13 07:22:22.573182 containerd[1920]: time="2025-08-13T07:22:22.573158514Z" level=info msg="StartContainer for \"4d92366739885ea706be665b5820656471dedf5a6c81ff21acb017d952ebafff\" returns successfully" Aug 13 07:22:23.522149 kubelet[3245]: I0813 07:22:23.522052 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-8dbb44b87-hjlnr" podStartSLOduration=1.970552676 podStartE2EDuration="5.522034768s" podCreationTimestamp="2025-08-13 07:22:18 +0000 UTC" firstStartedPulling="2025-08-13 07:22:18.95920895 +0000 UTC m=+30.682170838" lastFinishedPulling="2025-08-13 07:22:22.510691044 +0000 UTC m=+34.233652930" observedRunningTime="2025-08-13 07:22:23.521540065 +0000 UTC m=+35.244501972" watchObservedRunningTime="2025-08-13 07:22:23.522034768 +0000 UTC m=+35.244996664" Aug 13 07:22:25.349896 containerd[1920]: time="2025-08-13T07:22:25.349829324Z" level=info msg="StopPodSandbox for \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\"" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.375 [INFO][5497] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.376 [INFO][5497] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" iface="eth0" netns="/var/run/netns/cni-082bae9b-2356-8043-3ca9-75c077b6b2ff" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.376 [INFO][5497] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" iface="eth0" netns="/var/run/netns/cni-082bae9b-2356-8043-3ca9-75c077b6b2ff" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.376 [INFO][5497] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" iface="eth0" netns="/var/run/netns/cni-082bae9b-2356-8043-3ca9-75c077b6b2ff" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.376 [INFO][5497] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.376 [INFO][5497] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.388 [INFO][5512] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.388 [INFO][5512] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.388 [INFO][5512] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.392 [WARNING][5512] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.393 [INFO][5512] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.393 [INFO][5512] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:25.395723 containerd[1920]: 2025-08-13 07:22:25.394 [INFO][5497] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:25.396188 containerd[1920]: time="2025-08-13T07:22:25.395815561Z" level=info msg="TearDown network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\" successfully" Aug 13 07:22:25.396188 containerd[1920]: time="2025-08-13T07:22:25.395837077Z" level=info msg="StopPodSandbox for \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\" returns successfully" Aug 13 07:22:25.396264 containerd[1920]: time="2025-08-13T07:22:25.396248858Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-brr5j,Uid:89af5e87-c375-4eee-977b-cf953f1e3e5b,Namespace:kube-system,Attempt:1,}" Aug 13 07:22:25.397607 systemd[1]: run-netns-cni\x2d082bae9b\x2d2356\x2d8043\x2d3ca9\x2d75c077b6b2ff.mount: Deactivated successfully. Aug 13 07:22:25.448455 systemd-networkd[1561]: cali6514525fc53: Link UP Aug 13 07:22:25.448574 systemd-networkd[1561]: cali6514525fc53: Gained carrier Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.412 [INFO][5572] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.419 [INFO][5572] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0 coredns-7c65d6cfc9- kube-system 89af5e87-c375-4eee-977b-cf953f1e3e5b 892 0 2025-08-13 07:21:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f coredns-7c65d6cfc9-brr5j eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali6514525fc53 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.419 [INFO][5572] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.430 [INFO][5594] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" HandleID="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.430 [INFO][5594] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" HandleID="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001a57a0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"coredns-7c65d6cfc9-brr5j", "timestamp":"2025-08-13 07:22:25.430770675 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.430 [INFO][5594] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.430 [INFO][5594] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.430 [INFO][5594] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.434 [INFO][5594] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.436 [INFO][5594] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.439 [INFO][5594] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.440 [INFO][5594] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.441 [INFO][5594] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.441 [INFO][5594] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.441 [INFO][5594] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.444 [INFO][5594] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.446 [INFO][5594] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.66/26] block=192.168.121.64/26 handle="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.446 [INFO][5594] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.66/26] handle="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.446 [INFO][5594] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:25.454183 containerd[1920]: 2025-08-13 07:22:25.446 [INFO][5594] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.66/26] IPv6=[] ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" HandleID="k8s-pod-network.87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.454570 containerd[1920]: 2025-08-13 07:22:25.447 [INFO][5572] cni-plugin/k8s.go 418: Populated endpoint ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"89af5e87-c375-4eee-977b-cf953f1e3e5b", ResourceVersion:"892", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"coredns-7c65d6cfc9-brr5j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6514525fc53", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:25.454570 containerd[1920]: 2025-08-13 07:22:25.447 [INFO][5572] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.66/32] ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.454570 containerd[1920]: 2025-08-13 07:22:25.447 [INFO][5572] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6514525fc53 ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.454570 containerd[1920]: 2025-08-13 07:22:25.448 [INFO][5572] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.454570 containerd[1920]: 2025-08-13 07:22:25.448 [INFO][5572] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"89af5e87-c375-4eee-977b-cf953f1e3e5b", ResourceVersion:"892", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b", Pod:"coredns-7c65d6cfc9-brr5j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6514525fc53", MAC:"3a:22:60:0b:e8:80", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:25.454570 containerd[1920]: 2025-08-13 07:22:25.453 [INFO][5572] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-brr5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:25.462848 containerd[1920]: time="2025-08-13T07:22:25.462800606Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:25.463035 containerd[1920]: time="2025-08-13T07:22:25.462839842Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:25.463062 containerd[1920]: time="2025-08-13T07:22:25.463036384Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:25.463107 containerd[1920]: time="2025-08-13T07:22:25.463086539Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:25.497683 containerd[1920]: time="2025-08-13T07:22:25.497634332Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-brr5j,Uid:89af5e87-c375-4eee-977b-cf953f1e3e5b,Namespace:kube-system,Attempt:1,} returns sandbox id \"87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b\"" Aug 13 07:22:25.498756 containerd[1920]: time="2025-08-13T07:22:25.498744395Z" level=info msg="CreateContainer within sandbox \"87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:22:25.503382 containerd[1920]: time="2025-08-13T07:22:25.503346532Z" level=info msg="CreateContainer within sandbox \"87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"57963992044356c8f2a72e9958177681e976008fe5bb362174f1f62299fb4e36\"" Aug 13 07:22:25.503538 containerd[1920]: time="2025-08-13T07:22:25.503524359Z" level=info msg="StartContainer for \"57963992044356c8f2a72e9958177681e976008fe5bb362174f1f62299fb4e36\"" Aug 13 07:22:25.585249 containerd[1920]: time="2025-08-13T07:22:25.585205689Z" level=info msg="StartContainer for \"57963992044356c8f2a72e9958177681e976008fe5bb362174f1f62299fb4e36\" returns successfully" Aug 13 07:22:26.349915 containerd[1920]: time="2025-08-13T07:22:26.349820544Z" level=info msg="StopPodSandbox for \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\"" Aug 13 07:22:26.351365 containerd[1920]: time="2025-08-13T07:22:26.350157830Z" level=info msg="StopPodSandbox for \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\"" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.406 [INFO][5731] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.406 [INFO][5731] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" iface="eth0" netns="/var/run/netns/cni-486ee26a-2a91-20ce-4415-d44d6f348e93" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.406 [INFO][5731] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" iface="eth0" netns="/var/run/netns/cni-486ee26a-2a91-20ce-4415-d44d6f348e93" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.406 [INFO][5731] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" iface="eth0" netns="/var/run/netns/cni-486ee26a-2a91-20ce-4415-d44d6f348e93" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5731] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5731] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.419 [INFO][5760] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.419 [INFO][5760] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.419 [INFO][5760] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.424 [WARNING][5760] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.424 [INFO][5760] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.424 [INFO][5760] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:26.426278 containerd[1920]: 2025-08-13 07:22:26.425 [INFO][5731] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:26.426618 containerd[1920]: time="2025-08-13T07:22:26.426381865Z" level=info msg="TearDown network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\" successfully" Aug 13 07:22:26.426618 containerd[1920]: time="2025-08-13T07:22:26.426403606Z" level=info msg="StopPodSandbox for \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\" returns successfully" Aug 13 07:22:26.426803 containerd[1920]: time="2025-08-13T07:22:26.426784873Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xnn8d,Uid:b90d45be-b93c-47a4-ad3d-74de560838fa,Namespace:calico-system,Attempt:1,}" Aug 13 07:22:26.429286 systemd[1]: run-netns-cni\x2d486ee26a\x2d2a91\x2d20ce\x2d4415\x2dd44d6f348e93.mount: Deactivated successfully. Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5732] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5732] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" iface="eth0" netns="/var/run/netns/cni-5fb86ed1-70e1-f569-cd74-53432fb34a00" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5732] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" iface="eth0" netns="/var/run/netns/cni-5fb86ed1-70e1-f569-cd74-53432fb34a00" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5732] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" iface="eth0" netns="/var/run/netns/cni-5fb86ed1-70e1-f569-cd74-53432fb34a00" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5732] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.407 [INFO][5732] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.419 [INFO][5762] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.419 [INFO][5762] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.424 [INFO][5762] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.429 [WARNING][5762] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.429 [INFO][5762] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.429 [INFO][5762] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:26.431577 containerd[1920]: 2025-08-13 07:22:26.430 [INFO][5732] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:26.431853 containerd[1920]: time="2025-08-13T07:22:26.431656567Z" level=info msg="TearDown network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\" successfully" Aug 13 07:22:26.431853 containerd[1920]: time="2025-08-13T07:22:26.431672329Z" level=info msg="StopPodSandbox for \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\" returns successfully" Aug 13 07:22:26.432114 containerd[1920]: time="2025-08-13T07:22:26.432089114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-ggl67,Uid:926b03e8-ac14-4763-b7d3-fa81c377f2dd,Namespace:kube-system,Attempt:1,}" Aug 13 07:22:26.434648 systemd[1]: run-netns-cni\x2d5fb86ed1\x2d70e1\x2df569\x2dcd74\x2d53432fb34a00.mount: Deactivated successfully. Aug 13 07:22:26.486086 kubelet[3245]: I0813 07:22:26.486048 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-brr5j" podStartSLOduration=32.486032845 podStartE2EDuration="32.486032845s" podCreationTimestamp="2025-08-13 07:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:22:26.485750545 +0000 UTC m=+38.208712431" watchObservedRunningTime="2025-08-13 07:22:26.486032845 +0000 UTC m=+38.208994727" Aug 13 07:22:26.495171 systemd-networkd[1561]: calif7fbbc48599: Link UP Aug 13 07:22:26.495311 systemd-networkd[1561]: calif7fbbc48599: Gained carrier Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.460 [INFO][5841] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.466 [INFO][5841] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0 csi-node-driver- calico-system b90d45be-b93c-47a4-ad3d-74de560838fa 903 0 2025-08-13 07:22:05 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:57bd658777 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f csi-node-driver-xnn8d eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calif7fbbc48599 [] [] }} ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.466 [INFO][5841] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5888] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" HandleID="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5888] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" HandleID="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004ee80), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"csi-node-driver-xnn8d", "timestamp":"2025-08-13 07:22:26.477770044 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5888] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5888] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5888] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.481 [INFO][5888] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.483 [INFO][5888] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.485 [INFO][5888] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.486 [INFO][5888] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.488 [INFO][5888] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.488 [INFO][5888] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.489 [INFO][5888] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.490 [INFO][5888] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.493 [INFO][5888] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.67/26] block=192.168.121.64/26 handle="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.493 [INFO][5888] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.67/26] handle="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.493 [INFO][5888] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:26.500716 containerd[1920]: 2025-08-13 07:22:26.493 [INFO][5888] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.67/26] IPv6=[] ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" HandleID="k8s-pod-network.0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.501182 containerd[1920]: 2025-08-13 07:22:26.494 [INFO][5841] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b90d45be-b93c-47a4-ad3d-74de560838fa", ResourceVersion:"903", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"csi-node-driver-xnn8d", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.121.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif7fbbc48599", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:26.501182 containerd[1920]: 2025-08-13 07:22:26.494 [INFO][5841] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.67/32] ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.501182 containerd[1920]: 2025-08-13 07:22:26.494 [INFO][5841] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif7fbbc48599 ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.501182 containerd[1920]: 2025-08-13 07:22:26.495 [INFO][5841] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.501182 containerd[1920]: 2025-08-13 07:22:26.495 [INFO][5841] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b90d45be-b93c-47a4-ad3d-74de560838fa", ResourceVersion:"903", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f", Pod:"csi-node-driver-xnn8d", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.121.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif7fbbc48599", MAC:"66:85:90:6f:b6:25", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:26.501182 containerd[1920]: 2025-08-13 07:22:26.499 [INFO][5841] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f" Namespace="calico-system" Pod="csi-node-driver-xnn8d" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:26.508685 containerd[1920]: time="2025-08-13T07:22:26.508647606Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:26.508685 containerd[1920]: time="2025-08-13T07:22:26.508679632Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:26.508685 containerd[1920]: time="2025-08-13T07:22:26.508687053Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:26.508797 containerd[1920]: time="2025-08-13T07:22:26.508734061Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:26.529507 containerd[1920]: time="2025-08-13T07:22:26.529486311Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-xnn8d,Uid:b90d45be-b93c-47a4-ad3d-74de560838fa,Namespace:calico-system,Attempt:1,} returns sandbox id \"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f\"" Aug 13 07:22:26.530143 containerd[1920]: time="2025-08-13T07:22:26.530131216Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Aug 13 07:22:26.640462 systemd-networkd[1561]: cali1b1de5ae811: Link UP Aug 13 07:22:26.641210 systemd-networkd[1561]: cali1b1de5ae811: Gained carrier Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.460 [INFO][5840] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.466 [INFO][5840] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0 coredns-7c65d6cfc9- kube-system 926b03e8-ac14-4763-b7d3-fa81c377f2dd 904 0 2025-08-13 07:21:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f coredns-7c65d6cfc9-ggl67 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali1b1de5ae811 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.466 [INFO][5840] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5886] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" HandleID="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5886] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" HandleID="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000138da0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"coredns-7c65d6cfc9-ggl67", "timestamp":"2025-08-13 07:22:26.477770746 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.477 [INFO][5886] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.493 [INFO][5886] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.493 [INFO][5886] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.584 [INFO][5886] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.593 [INFO][5886] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.602 [INFO][5886] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.606 [INFO][5886] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.611 [INFO][5886] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.611 [INFO][5886] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.614 [INFO][5886] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123 Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.621 [INFO][5886] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.631 [INFO][5886] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.68/26] block=192.168.121.64/26 handle="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.632 [INFO][5886] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.68/26] handle="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.632 [INFO][5886] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:26.668955 containerd[1920]: 2025-08-13 07:22:26.632 [INFO][5886] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.68/26] IPv6=[] ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" HandleID="k8s-pod-network.9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.671504 containerd[1920]: 2025-08-13 07:22:26.636 [INFO][5840] cni-plugin/k8s.go 418: Populated endpoint ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"926b03e8-ac14-4763-b7d3-fa81c377f2dd", ResourceVersion:"904", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"coredns-7c65d6cfc9-ggl67", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1b1de5ae811", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:26.671504 containerd[1920]: 2025-08-13 07:22:26.636 [INFO][5840] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.68/32] ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.671504 containerd[1920]: 2025-08-13 07:22:26.637 [INFO][5840] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1b1de5ae811 ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.671504 containerd[1920]: 2025-08-13 07:22:26.641 [INFO][5840] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.671504 containerd[1920]: 2025-08-13 07:22:26.641 [INFO][5840] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"926b03e8-ac14-4763-b7d3-fa81c377f2dd", ResourceVersion:"904", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123", Pod:"coredns-7c65d6cfc9-ggl67", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1b1de5ae811", MAC:"ce:e2:2e:73:3b:07", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:26.671504 containerd[1920]: 2025-08-13 07:22:26.663 [INFO][5840] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123" Namespace="kube-system" Pod="coredns-7c65d6cfc9-ggl67" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:26.682776 containerd[1920]: time="2025-08-13T07:22:26.682691175Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:26.682776 containerd[1920]: time="2025-08-13T07:22:26.682721356Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:26.682776 containerd[1920]: time="2025-08-13T07:22:26.682728344Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:26.682892 containerd[1920]: time="2025-08-13T07:22:26.682795544Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:26.784740 containerd[1920]: time="2025-08-13T07:22:26.784711639Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-ggl67,Uid:926b03e8-ac14-4763-b7d3-fa81c377f2dd,Namespace:kube-system,Attempt:1,} returns sandbox id \"9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123\"" Aug 13 07:22:26.786127 containerd[1920]: time="2025-08-13T07:22:26.786108869Z" level=info msg="CreateContainer within sandbox \"9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:22:26.790082 containerd[1920]: time="2025-08-13T07:22:26.790069039Z" level=info msg="CreateContainer within sandbox \"9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"dc3e32239f40e1e5f779cdd7a75aa34c5d02c28086be9fee1a1f1300d840dc3d\"" Aug 13 07:22:26.790306 containerd[1920]: time="2025-08-13T07:22:26.790293386Z" level=info msg="StartContainer for \"dc3e32239f40e1e5f779cdd7a75aa34c5d02c28086be9fee1a1f1300d840dc3d\"" Aug 13 07:22:26.795168 systemd-networkd[1561]: cali6514525fc53: Gained IPv6LL Aug 13 07:22:26.830272 containerd[1920]: time="2025-08-13T07:22:26.830225378Z" level=info msg="StartContainer for \"dc3e32239f40e1e5f779cdd7a75aa34c5d02c28086be9fee1a1f1300d840dc3d\" returns successfully" Aug 13 07:22:27.350621 containerd[1920]: time="2025-08-13T07:22:27.350515350Z" level=info msg="StopPodSandbox for \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\"" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.384 [INFO][6078] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.384 [INFO][6078] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" iface="eth0" netns="/var/run/netns/cni-ebb3285a-de33-c143-a421-519d1b0287c3" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.384 [INFO][6078] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" iface="eth0" netns="/var/run/netns/cni-ebb3285a-de33-c143-a421-519d1b0287c3" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.384 [INFO][6078] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" iface="eth0" netns="/var/run/netns/cni-ebb3285a-de33-c143-a421-519d1b0287c3" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.384 [INFO][6078] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.384 [INFO][6078] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.394 [INFO][6095] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.394 [INFO][6095] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.394 [INFO][6095] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.398 [WARNING][6095] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.398 [INFO][6095] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.399 [INFO][6095] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:27.400868 containerd[1920]: 2025-08-13 07:22:27.400 [INFO][6078] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:27.401191 containerd[1920]: time="2025-08-13T07:22:27.400912000Z" level=info msg="TearDown network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\" successfully" Aug 13 07:22:27.401191 containerd[1920]: time="2025-08-13T07:22:27.400929972Z" level=info msg="StopPodSandbox for \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\" returns successfully" Aug 13 07:22:27.401348 containerd[1920]: time="2025-08-13T07:22:27.401307495Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-qqszn,Uid:bb4a7045-3301-41f3-86a2-df653630104c,Namespace:calico-apiserver,Attempt:1,}" Aug 13 07:22:27.402487 systemd[1]: run-netns-cni\x2debb3285a\x2dde33\x2dc143\x2da421\x2d519d1b0287c3.mount: Deactivated successfully. Aug 13 07:22:27.482427 systemd-networkd[1561]: cali5327790de75: Link UP Aug 13 07:22:27.482565 systemd-networkd[1561]: cali5327790de75: Gained carrier Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.419 [INFO][6108] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.425 [INFO][6108] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0 calico-apiserver-59fbd88f7c- calico-apiserver bb4a7045-3301-41f3-86a2-df653630104c 927 0 2025-08-13 07:22:01 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:59fbd88f7c projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f calico-apiserver-59fbd88f7c-qqszn eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali5327790de75 [] [] }} ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.425 [INFO][6108] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.437 [INFO][6130] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" HandleID="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.437 [INFO][6130] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" HandleID="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000590fd0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"calico-apiserver-59fbd88f7c-qqszn", "timestamp":"2025-08-13 07:22:27.437540367 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.437 [INFO][6130] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.437 [INFO][6130] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.437 [INFO][6130] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.441 [INFO][6130] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.470 [INFO][6130] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.472 [INFO][6130] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.473 [INFO][6130] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.474 [INFO][6130] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.474 [INFO][6130] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.475 [INFO][6130] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.477 [INFO][6130] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.480 [INFO][6130] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.69/26] block=192.168.121.64/26 handle="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.480 [INFO][6130] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.69/26] handle="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.480 [INFO][6130] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:27.487663 containerd[1920]: 2025-08-13 07:22:27.480 [INFO][6130] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.69/26] IPv6=[] ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" HandleID="k8s-pod-network.abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.488156 containerd[1920]: 2025-08-13 07:22:27.481 [INFO][6108] cni-plugin/k8s.go 418: Populated endpoint ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"bb4a7045-3301-41f3-86a2-df653630104c", ResourceVersion:"927", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"calico-apiserver-59fbd88f7c-qqszn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5327790de75", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:27.488156 containerd[1920]: 2025-08-13 07:22:27.481 [INFO][6108] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.69/32] ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.488156 containerd[1920]: 2025-08-13 07:22:27.481 [INFO][6108] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5327790de75 ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.488156 containerd[1920]: 2025-08-13 07:22:27.482 [INFO][6108] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.488156 containerd[1920]: 2025-08-13 07:22:27.482 [INFO][6108] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"bb4a7045-3301-41f3-86a2-df653630104c", ResourceVersion:"927", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b", Pod:"calico-apiserver-59fbd88f7c-qqszn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5327790de75", MAC:"72:30:3c:21:24:04", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:27.488156 containerd[1920]: 2025-08-13 07:22:27.486 [INFO][6108] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-qqszn" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:27.489178 kubelet[3245]: I0813 07:22:27.489140 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-ggl67" podStartSLOduration=33.489121883 podStartE2EDuration="33.489121883s" podCreationTimestamp="2025-08-13 07:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:22:27.489116453 +0000 UTC m=+39.212078340" watchObservedRunningTime="2025-08-13 07:22:27.489121883 +0000 UTC m=+39.212083774" Aug 13 07:22:27.495983 containerd[1920]: time="2025-08-13T07:22:27.495915230Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:27.495983 containerd[1920]: time="2025-08-13T07:22:27.495945644Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:27.496192 containerd[1920]: time="2025-08-13T07:22:27.495965128Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:27.496277 containerd[1920]: time="2025-08-13T07:22:27.496202071Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:27.530422 containerd[1920]: time="2025-08-13T07:22:27.530369104Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-qqszn,Uid:bb4a7045-3301-41f3-86a2-df653630104c,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b\"" Aug 13 07:22:27.690607 systemd-networkd[1561]: cali1b1de5ae811: Gained IPv6LL Aug 13 07:22:28.000408 containerd[1920]: time="2025-08-13T07:22:28.000346465Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:28.000578 containerd[1920]: time="2025-08-13T07:22:28.000534228Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.2: active requests=0, bytes read=8759190" Aug 13 07:22:28.000952 containerd[1920]: time="2025-08-13T07:22:28.000916427Z" level=info msg="ImageCreate event name:\"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:28.001926 containerd[1920]: time="2025-08-13T07:22:28.001891075Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:28.002470 containerd[1920]: time="2025-08-13T07:22:28.002429502Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.2\" with image id \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\", size \"10251893\" in 1.472279705s" Aug 13 07:22:28.002470 containerd[1920]: time="2025-08-13T07:22:28.002451751Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\"" Aug 13 07:22:28.002991 containerd[1920]: time="2025-08-13T07:22:28.002957764Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 07:22:28.003714 containerd[1920]: time="2025-08-13T07:22:28.003673037Z" level=info msg="CreateContainer within sandbox \"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Aug 13 07:22:28.009047 containerd[1920]: time="2025-08-13T07:22:28.009007110Z" level=info msg="CreateContainer within sandbox \"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"2780a26f0f1c4d4d5cb45552a7948e8218c468d2e5818b77101cefd7b138983b\"" Aug 13 07:22:28.009313 containerd[1920]: time="2025-08-13T07:22:28.009271055Z" level=info msg="StartContainer for \"2780a26f0f1c4d4d5cb45552a7948e8218c468d2e5818b77101cefd7b138983b\"" Aug 13 07:22:28.010207 systemd-networkd[1561]: calif7fbbc48599: Gained IPv6LL Aug 13 07:22:28.051607 containerd[1920]: time="2025-08-13T07:22:28.051559180Z" level=info msg="StartContainer for \"2780a26f0f1c4d4d5cb45552a7948e8218c468d2e5818b77101cefd7b138983b\" returns successfully" Aug 13 07:22:28.349157 containerd[1920]: time="2025-08-13T07:22:28.349100979Z" level=info msg="StopPodSandbox for \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\"" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.373 [INFO][6296] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.374 [INFO][6296] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" iface="eth0" netns="/var/run/netns/cni-393775b6-4938-f605-1fbf-a2a8abb938bb" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.374 [INFO][6296] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" iface="eth0" netns="/var/run/netns/cni-393775b6-4938-f605-1fbf-a2a8abb938bb" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.374 [INFO][6296] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" iface="eth0" netns="/var/run/netns/cni-393775b6-4938-f605-1fbf-a2a8abb938bb" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.374 [INFO][6296] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.374 [INFO][6296] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.391 [INFO][6314] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.391 [INFO][6314] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.391 [INFO][6314] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.398 [WARNING][6314] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.398 [INFO][6314] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.400 [INFO][6314] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:28.402804 containerd[1920]: 2025-08-13 07:22:28.401 [INFO][6296] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:28.403518 containerd[1920]: time="2025-08-13T07:22:28.402932400Z" level=info msg="TearDown network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\" successfully" Aug 13 07:22:28.403518 containerd[1920]: time="2025-08-13T07:22:28.402957461Z" level=info msg="StopPodSandbox for \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\" returns successfully" Aug 13 07:22:28.403635 containerd[1920]: time="2025-08-13T07:22:28.403577332Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-694d7cc5f7-rl69w,Uid:f80e5fda-a5d5-4f6a-9768-261c847bf3ee,Namespace:calico-system,Attempt:1,}" Aug 13 07:22:28.405343 systemd[1]: run-netns-cni\x2d393775b6\x2d4938\x2df605\x2d1fbf\x2da2a8abb938bb.mount: Deactivated successfully. Aug 13 07:22:28.458473 systemd-networkd[1561]: calie575646a156: Link UP Aug 13 07:22:28.458648 systemd-networkd[1561]: calie575646a156: Gained carrier Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.417 [INFO][6331] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.424 [INFO][6331] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0 calico-kube-controllers-694d7cc5f7- calico-system f80e5fda-a5d5-4f6a-9768-261c847bf3ee 941 0 2025-08-13 07:22:05 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:694d7cc5f7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f calico-kube-controllers-694d7cc5f7-rl69w eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calie575646a156 [] [] }} ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.424 [INFO][6331] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.435 [INFO][6348] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" HandleID="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.435 [INFO][6348] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" HandleID="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fbf0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"calico-kube-controllers-694d7cc5f7-rl69w", "timestamp":"2025-08-13 07:22:28.435855865 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.435 [INFO][6348] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.435 [INFO][6348] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.436 [INFO][6348] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.440 [INFO][6348] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.443 [INFO][6348] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.446 [INFO][6348] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.448 [INFO][6348] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.449 [INFO][6348] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.449 [INFO][6348] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.450 [INFO][6348] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42 Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.452 [INFO][6348] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.456 [INFO][6348] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.70/26] block=192.168.121.64/26 handle="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.456 [INFO][6348] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.70/26] handle="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.456 [INFO][6348] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:28.466283 containerd[1920]: 2025-08-13 07:22:28.456 [INFO][6348] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.70/26] IPv6=[] ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" HandleID="k8s-pod-network.da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.466716 containerd[1920]: 2025-08-13 07:22:28.457 [INFO][6331] cni-plugin/k8s.go 418: Populated endpoint ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0", GenerateName:"calico-kube-controllers-694d7cc5f7-", Namespace:"calico-system", SelfLink:"", UID:"f80e5fda-a5d5-4f6a-9768-261c847bf3ee", ResourceVersion:"941", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"694d7cc5f7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"calico-kube-controllers-694d7cc5f7-rl69w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.121.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie575646a156", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:28.466716 containerd[1920]: 2025-08-13 07:22:28.457 [INFO][6331] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.70/32] ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.466716 containerd[1920]: 2025-08-13 07:22:28.457 [INFO][6331] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie575646a156 ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.466716 containerd[1920]: 2025-08-13 07:22:28.459 [INFO][6331] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.466716 containerd[1920]: 2025-08-13 07:22:28.459 [INFO][6331] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0", GenerateName:"calico-kube-controllers-694d7cc5f7-", Namespace:"calico-system", SelfLink:"", UID:"f80e5fda-a5d5-4f6a-9768-261c847bf3ee", ResourceVersion:"941", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"694d7cc5f7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42", Pod:"calico-kube-controllers-694d7cc5f7-rl69w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.121.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie575646a156", MAC:"d2:3d:db:f4:3e:5d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:28.466716 containerd[1920]: 2025-08-13 07:22:28.465 [INFO][6331] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42" Namespace="calico-system" Pod="calico-kube-controllers-694d7cc5f7-rl69w" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:28.474346 containerd[1920]: time="2025-08-13T07:22:28.474278011Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:28.474560 containerd[1920]: time="2025-08-13T07:22:28.474330912Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:28.474560 containerd[1920]: time="2025-08-13T07:22:28.474534100Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:28.474654 containerd[1920]: time="2025-08-13T07:22:28.474614530Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:28.513961 containerd[1920]: time="2025-08-13T07:22:28.513939468Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-694d7cc5f7-rl69w,Uid:f80e5fda-a5d5-4f6a-9768-261c847bf3ee,Namespace:calico-system,Attempt:1,} returns sandbox id \"da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42\"" Aug 13 07:22:29.034450 systemd-networkd[1561]: cali5327790de75: Gained IPv6LL Aug 13 07:22:29.349762 containerd[1920]: time="2025-08-13T07:22:29.349627231Z" level=info msg="StopPodSandbox for \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\"" Aug 13 07:22:29.349762 containerd[1920]: time="2025-08-13T07:22:29.349728900Z" level=info msg="StopPodSandbox for \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\"" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6490] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6490] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" iface="eth0" netns="/var/run/netns/cni-982d71cd-a5c1-3fcf-3d5a-8a4d898c9880" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6490] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" iface="eth0" netns="/var/run/netns/cni-982d71cd-a5c1-3fcf-3d5a-8a4d898c9880" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6490] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" iface="eth0" netns="/var/run/netns/cni-982d71cd-a5c1-3fcf-3d5a-8a4d898c9880" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6490] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6490] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.381 [INFO][6523] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.381 [INFO][6523] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.381 [INFO][6523] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.384 [WARNING][6523] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.384 [INFO][6523] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.385 [INFO][6523] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:29.387008 containerd[1920]: 2025-08-13 07:22:29.386 [INFO][6490] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:29.387292 containerd[1920]: time="2025-08-13T07:22:29.387071877Z" level=info msg="TearDown network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\" successfully" Aug 13 07:22:29.387292 containerd[1920]: time="2025-08-13T07:22:29.387088616Z" level=info msg="StopPodSandbox for \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\" returns successfully" Aug 13 07:22:29.387461 containerd[1920]: time="2025-08-13T07:22:29.387424863Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-z2k5j,Uid:b2d55415-dd30-4e5a-9264-00bc2660bdba,Namespace:calico-apiserver,Attempt:1,}" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6489] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6489] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" iface="eth0" netns="/var/run/netns/cni-51f20872-1f64-0264-2229-42328354bbba" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6489] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" iface="eth0" netns="/var/run/netns/cni-51f20872-1f64-0264-2229-42328354bbba" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6489] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" iface="eth0" netns="/var/run/netns/cni-51f20872-1f64-0264-2229-42328354bbba" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6489] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.371 [INFO][6489] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.381 [INFO][6525] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.381 [INFO][6525] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.385 [INFO][6525] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.388 [WARNING][6525] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.388 [INFO][6525] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.389 [INFO][6525] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:29.391191 containerd[1920]: 2025-08-13 07:22:29.390 [INFO][6489] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:29.391438 containerd[1920]: time="2025-08-13T07:22:29.391276783Z" level=info msg="TearDown network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\" successfully" Aug 13 07:22:29.391438 containerd[1920]: time="2025-08-13T07:22:29.391291909Z" level=info msg="StopPodSandbox for \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\" returns successfully" Aug 13 07:22:29.391629 containerd[1920]: time="2025-08-13T07:22:29.391617379Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-sjk28,Uid:dd85a865-2c7f-4339-b9a7-90c5b37bd6d6,Namespace:calico-system,Attempt:1,}" Aug 13 07:22:29.401935 systemd[1]: run-netns-cni\x2d51f20872\x2d1f64\x2d0264\x2d2229\x2d42328354bbba.mount: Deactivated successfully. Aug 13 07:22:29.402052 systemd[1]: run-netns-cni\x2d982d71cd\x2da5c1\x2d3fcf\x2d3d5a\x2d8a4d898c9880.mount: Deactivated successfully. Aug 13 07:22:29.446026 systemd-networkd[1561]: calia774e84ab68: Link UP Aug 13 07:22:29.446170 systemd-networkd[1561]: calia774e84ab68: Gained carrier Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.402 [INFO][6556] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.411 [INFO][6556] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0 calico-apiserver-59fbd88f7c- calico-apiserver b2d55415-dd30-4e5a-9264-00bc2660bdba 955 0 2025-08-13 07:22:01 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:59fbd88f7c projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f calico-apiserver-59fbd88f7c-z2k5j eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calia774e84ab68 [] [] }} ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.411 [INFO][6556] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.425 [INFO][6604] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" HandleID="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.425 [INFO][6604] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" HandleID="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000138b50), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"calico-apiserver-59fbd88f7c-z2k5j", "timestamp":"2025-08-13 07:22:29.425317729 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.425 [INFO][6604] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.425 [INFO][6604] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.425 [INFO][6604] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.429 [INFO][6604] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.432 [INFO][6604] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.435 [INFO][6604] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.436 [INFO][6604] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.437 [INFO][6604] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.437 [INFO][6604] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.439 [INFO][6604] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7 Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.441 [INFO][6604] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.444 [INFO][6604] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.71/26] block=192.168.121.64/26 handle="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.444 [INFO][6604] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.71/26] handle="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.444 [INFO][6604] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:29.451813 containerd[1920]: 2025-08-13 07:22:29.444 [INFO][6604] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.71/26] IPv6=[] ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" HandleID="k8s-pod-network.558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.452355 containerd[1920]: 2025-08-13 07:22:29.445 [INFO][6556] cni-plugin/k8s.go 418: Populated endpoint ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"b2d55415-dd30-4e5a-9264-00bc2660bdba", ResourceVersion:"955", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"calico-apiserver-59fbd88f7c-z2k5j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia774e84ab68", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:29.452355 containerd[1920]: 2025-08-13 07:22:29.445 [INFO][6556] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.71/32] ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.452355 containerd[1920]: 2025-08-13 07:22:29.445 [INFO][6556] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia774e84ab68 ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.452355 containerd[1920]: 2025-08-13 07:22:29.446 [INFO][6556] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.452355 containerd[1920]: 2025-08-13 07:22:29.446 [INFO][6556] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"b2d55415-dd30-4e5a-9264-00bc2660bdba", ResourceVersion:"955", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7", Pod:"calico-apiserver-59fbd88f7c-z2k5j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia774e84ab68", MAC:"e6:26:e3:fc:fd:23", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:29.452355 containerd[1920]: 2025-08-13 07:22:29.451 [INFO][6556] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7" Namespace="calico-apiserver" Pod="calico-apiserver-59fbd88f7c-z2k5j" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:29.459984 containerd[1920]: time="2025-08-13T07:22:29.459943869Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:29.459984 containerd[1920]: time="2025-08-13T07:22:29.459974289Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:29.459984 containerd[1920]: time="2025-08-13T07:22:29.459981334Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:29.460138 containerd[1920]: time="2025-08-13T07:22:29.460028816Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:29.507242 containerd[1920]: time="2025-08-13T07:22:29.507216650Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-59fbd88f7c-z2k5j,Uid:b2d55415-dd30-4e5a-9264-00bc2660bdba,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7\"" Aug 13 07:22:29.552015 systemd-networkd[1561]: calicf213437f46: Link UP Aug 13 07:22:29.552205 systemd-networkd[1561]: calicf213437f46: Gained carrier Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.407 [INFO][6567] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.413 [INFO][6567] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0 goldmane-58fd7646b9- calico-system dd85a865-2c7f-4339-b9a7-90c5b37bd6d6 956 0 2025-08-13 07:22:04 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:58fd7646b9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081.3.5-a-3ca3f1276f goldmane-58fd7646b9-sjk28 eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calicf213437f46 [] [] }} ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.413 [INFO][6567] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.426 [INFO][6610] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" HandleID="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.426 [INFO][6610] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" HandleID="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003cf320), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-3ca3f1276f", "pod":"goldmane-58fd7646b9-sjk28", "timestamp":"2025-08-13 07:22:29.426119161 +0000 UTC"}, Hostname:"ci-4081.3.5-a-3ca3f1276f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.426 [INFO][6610] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.444 [INFO][6610] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.444 [INFO][6610] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-3ca3f1276f' Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.532 [INFO][6610] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.537 [INFO][6610] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.540 [INFO][6610] ipam/ipam.go 511: Trying affinity for 192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.542 [INFO][6610] ipam/ipam.go 158: Attempting to load block cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.543 [INFO][6610] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.121.64/26 host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.543 [INFO][6610] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.121.64/26 handle="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.544 [INFO][6610] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765 Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.546 [INFO][6610] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.121.64/26 handle="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.550 [INFO][6610] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.121.72/26] block=192.168.121.64/26 handle="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.550 [INFO][6610] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.121.72/26] handle="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" host="ci-4081.3.5-a-3ca3f1276f" Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.550 [INFO][6610] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:29.557514 containerd[1920]: 2025-08-13 07:22:29.550 [INFO][6610] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.121.72/26] IPv6=[] ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" HandleID="k8s-pod-network.24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.557904 containerd[1920]: 2025-08-13 07:22:29.551 [INFO][6567] cni-plugin/k8s.go 418: Populated endpoint ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6", ResourceVersion:"956", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"", Pod:"goldmane-58fd7646b9-sjk28", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.121.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicf213437f46", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:29.557904 containerd[1920]: 2025-08-13 07:22:29.551 [INFO][6567] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.121.72/32] ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.557904 containerd[1920]: 2025-08-13 07:22:29.551 [INFO][6567] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicf213437f46 ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.557904 containerd[1920]: 2025-08-13 07:22:29.552 [INFO][6567] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.557904 containerd[1920]: 2025-08-13 07:22:29.552 [INFO][6567] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6", ResourceVersion:"956", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765", Pod:"goldmane-58fd7646b9-sjk28", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.121.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicf213437f46", MAC:"86:5d:37:aa:6d:b9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:29.557904 containerd[1920]: 2025-08-13 07:22:29.556 [INFO][6567] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765" Namespace="calico-system" Pod="goldmane-58fd7646b9-sjk28" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:29.565280 containerd[1920]: time="2025-08-13T07:22:29.565232526Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:22:29.565515 containerd[1920]: time="2025-08-13T07:22:29.565279326Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:22:29.565550 containerd[1920]: time="2025-08-13T07:22:29.565531619Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:29.565615 containerd[1920]: time="2025-08-13T07:22:29.565604151Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:22:29.600354 containerd[1920]: time="2025-08-13T07:22:29.600295030Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-sjk28,Uid:dd85a865-2c7f-4339-b9a7-90c5b37bd6d6,Namespace:calico-system,Attempt:1,} returns sandbox id \"24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765\"" Aug 13 07:22:29.723532 kubelet[3245]: I0813 07:22:29.723503 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:30.003047 containerd[1920]: time="2025-08-13T07:22:30.002965540Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:30.003277 containerd[1920]: time="2025-08-13T07:22:30.003227781Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=47317977" Aug 13 07:22:30.003608 containerd[1920]: time="2025-08-13T07:22:30.003568825Z" level=info msg="ImageCreate event name:\"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:30.004598 containerd[1920]: time="2025-08-13T07:22:30.004556193Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:30.005054 containerd[1920]: time="2025-08-13T07:22:30.005012264Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 2.00203676s" Aug 13 07:22:30.005054 containerd[1920]: time="2025-08-13T07:22:30.005030054Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Aug 13 07:22:30.005614 containerd[1920]: time="2025-08-13T07:22:30.005578118Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Aug 13 07:22:30.005993 containerd[1920]: time="2025-08-13T07:22:30.005980685Z" level=info msg="CreateContainer within sandbox \"abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 07:22:30.009314 containerd[1920]: time="2025-08-13T07:22:30.009272791Z" level=info msg="CreateContainer within sandbox \"abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"47731b5a9d78882c2f34e5f98cafecb9272a349f7510e12f49428492b6dd9b70\"" Aug 13 07:22:30.009563 containerd[1920]: time="2025-08-13T07:22:30.009510576Z" level=info msg="StartContainer for \"47731b5a9d78882c2f34e5f98cafecb9272a349f7510e12f49428492b6dd9b70\"" Aug 13 07:22:30.061029 containerd[1920]: time="2025-08-13T07:22:30.060963071Z" level=info msg="StartContainer for \"47731b5a9d78882c2f34e5f98cafecb9272a349f7510e12f49428492b6dd9b70\" returns successfully" Aug 13 07:22:30.250196 systemd-networkd[1561]: calie575646a156: Gained IPv6LL Aug 13 07:22:30.469079 kernel: bpftool[6865]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Aug 13 07:22:30.499868 kubelet[3245]: I0813 07:22:30.499835 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-59fbd88f7c-qqszn" podStartSLOduration=27.025304915 podStartE2EDuration="29.499823786s" podCreationTimestamp="2025-08-13 07:22:01 +0000 UTC" firstStartedPulling="2025-08-13 07:22:27.530906479 +0000 UTC m=+39.253868363" lastFinishedPulling="2025-08-13 07:22:30.005425349 +0000 UTC m=+41.728387234" observedRunningTime="2025-08-13 07:22:30.499748763 +0000 UTC m=+42.222710653" watchObservedRunningTime="2025-08-13 07:22:30.499823786 +0000 UTC m=+42.222785668" Aug 13 07:22:30.529939 kubelet[3245]: I0813 07:22:30.529921 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:30.638088 systemd-networkd[1561]: vxlan.calico: Link UP Aug 13 07:22:30.638091 systemd-networkd[1561]: vxlan.calico: Gained carrier Aug 13 07:22:31.146422 systemd-networkd[1561]: calicf213437f46: Gained IPv6LL Aug 13 07:22:31.338331 systemd-networkd[1561]: calia774e84ab68: Gained IPv6LL Aug 13 07:22:31.494583 kubelet[3245]: I0813 07:22:31.494425 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:31.771587 containerd[1920]: time="2025-08-13T07:22:31.771499945Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:31.771810 containerd[1920]: time="2025-08-13T07:22:31.771726419Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2: active requests=0, bytes read=14703784" Aug 13 07:22:31.772115 containerd[1920]: time="2025-08-13T07:22:31.772085180Z" level=info msg="ImageCreate event name:\"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:31.773062 containerd[1920]: time="2025-08-13T07:22:31.773021197Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:31.773476 containerd[1920]: time="2025-08-13T07:22:31.773433495Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" with image id \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\", size \"16196439\" in 1.767840972s" Aug 13 07:22:31.773476 containerd[1920]: time="2025-08-13T07:22:31.773449800Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\"" Aug 13 07:22:31.773907 containerd[1920]: time="2025-08-13T07:22:31.773895553Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Aug 13 07:22:31.774500 containerd[1920]: time="2025-08-13T07:22:31.774487168Z" level=info msg="CreateContainer within sandbox \"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Aug 13 07:22:31.779084 containerd[1920]: time="2025-08-13T07:22:31.779037268Z" level=info msg="CreateContainer within sandbox \"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"a19d1ba506d49dfb01b1dd559b60c90bc20c07068fef63564eae990d8b1e4ba5\"" Aug 13 07:22:31.779335 containerd[1920]: time="2025-08-13T07:22:31.779278367Z" level=info msg="StartContainer for \"a19d1ba506d49dfb01b1dd559b60c90bc20c07068fef63564eae990d8b1e4ba5\"" Aug 13 07:22:31.820608 containerd[1920]: time="2025-08-13T07:22:31.820563648Z" level=info msg="StartContainer for \"a19d1ba506d49dfb01b1dd559b60c90bc20c07068fef63564eae990d8b1e4ba5\" returns successfully" Aug 13 07:22:31.914284 systemd-networkd[1561]: vxlan.calico: Gained IPv6LL Aug 13 07:22:32.396511 kubelet[3245]: I0813 07:22:32.396441 3245 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Aug 13 07:22:32.396511 kubelet[3245]: I0813 07:22:32.396535 3245 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Aug 13 07:22:32.509493 kubelet[3245]: I0813 07:22:32.509379 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-xnn8d" podStartSLOduration=22.26546296 podStartE2EDuration="27.509332305s" podCreationTimestamp="2025-08-13 07:22:05 +0000 UTC" firstStartedPulling="2025-08-13 07:22:26.529982737 +0000 UTC m=+38.252944621" lastFinishedPulling="2025-08-13 07:22:31.773852081 +0000 UTC m=+43.496813966" observedRunningTime="2025-08-13 07:22:32.50745539 +0000 UTC m=+44.230417278" watchObservedRunningTime="2025-08-13 07:22:32.509332305 +0000 UTC m=+44.232294239" Aug 13 07:22:34.085542 containerd[1920]: time="2025-08-13T07:22:34.085483573Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:34.085782 containerd[1920]: time="2025-08-13T07:22:34.085716686Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.2: active requests=0, bytes read=51276688" Aug 13 07:22:34.086144 containerd[1920]: time="2025-08-13T07:22:34.086106278Z" level=info msg="ImageCreate event name:\"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:34.087085 containerd[1920]: time="2025-08-13T07:22:34.087040215Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:34.087443 containerd[1920]: time="2025-08-13T07:22:34.087404209Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" with image id \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\", size \"52769359\" in 2.313494585s" Aug 13 07:22:34.087443 containerd[1920]: time="2025-08-13T07:22:34.087419042Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\"" Aug 13 07:22:34.087889 containerd[1920]: time="2025-08-13T07:22:34.087877643Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 07:22:34.090750 containerd[1920]: time="2025-08-13T07:22:34.090711715Z" level=info msg="CreateContainer within sandbox \"da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Aug 13 07:22:34.094328 containerd[1920]: time="2025-08-13T07:22:34.094285574Z" level=info msg="CreateContainer within sandbox \"da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"57a996d4f341499be585d6b7d1d7b9db4e746b13aeabcf872936ce14541e8c62\"" Aug 13 07:22:34.094583 containerd[1920]: time="2025-08-13T07:22:34.094571313Z" level=info msg="StartContainer for \"57a996d4f341499be585d6b7d1d7b9db4e746b13aeabcf872936ce14541e8c62\"" Aug 13 07:22:34.171822 containerd[1920]: time="2025-08-13T07:22:34.171799978Z" level=info msg="StartContainer for \"57a996d4f341499be585d6b7d1d7b9db4e746b13aeabcf872936ce14541e8c62\" returns successfully" Aug 13 07:22:34.533658 kubelet[3245]: I0813 07:22:34.533517 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-694d7cc5f7-rl69w" podStartSLOduration=23.960153645 podStartE2EDuration="29.533472695s" podCreationTimestamp="2025-08-13 07:22:05 +0000 UTC" firstStartedPulling="2025-08-13 07:22:28.514504366 +0000 UTC m=+40.237466251" lastFinishedPulling="2025-08-13 07:22:34.087823416 +0000 UTC m=+45.810785301" observedRunningTime="2025-08-13 07:22:34.532800202 +0000 UTC m=+46.255762183" watchObservedRunningTime="2025-08-13 07:22:34.533472695 +0000 UTC m=+46.256434631" Aug 13 07:22:34.968942 containerd[1920]: time="2025-08-13T07:22:34.968883400Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:34.969144 containerd[1920]: time="2025-08-13T07:22:34.969090721Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=77" Aug 13 07:22:34.970410 containerd[1920]: time="2025-08-13T07:22:34.970368930Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 882.475487ms" Aug 13 07:22:34.970410 containerd[1920]: time="2025-08-13T07:22:34.970384189Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Aug 13 07:22:34.970978 containerd[1920]: time="2025-08-13T07:22:34.970968525Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Aug 13 07:22:34.971616 containerd[1920]: time="2025-08-13T07:22:34.971600934Z" level=info msg="CreateContainer within sandbox \"558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 07:22:34.975671 containerd[1920]: time="2025-08-13T07:22:34.975627814Z" level=info msg="CreateContainer within sandbox \"558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"56f402c1cee7dc57b34c7fbfad01dad511a3530b9073da9937dbc90f2af663a6\"" Aug 13 07:22:34.975936 containerd[1920]: time="2025-08-13T07:22:34.975900874Z" level=info msg="StartContainer for \"56f402c1cee7dc57b34c7fbfad01dad511a3530b9073da9937dbc90f2af663a6\"" Aug 13 07:22:35.018903 containerd[1920]: time="2025-08-13T07:22:35.018857764Z" level=info msg="StartContainer for \"56f402c1cee7dc57b34c7fbfad01dad511a3530b9073da9937dbc90f2af663a6\" returns successfully" Aug 13 07:22:35.541451 kubelet[3245]: I0813 07:22:35.541347 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-59fbd88f7c-z2k5j" podStartSLOduration=29.078227744 podStartE2EDuration="34.541307273s" podCreationTimestamp="2025-08-13 07:22:01 +0000 UTC" firstStartedPulling="2025-08-13 07:22:29.507785016 +0000 UTC m=+41.230746901" lastFinishedPulling="2025-08-13 07:22:34.970864544 +0000 UTC m=+46.693826430" observedRunningTime="2025-08-13 07:22:35.540387039 +0000 UTC m=+47.263349021" watchObservedRunningTime="2025-08-13 07:22:35.541307273 +0000 UTC m=+47.264269206" Aug 13 07:22:35.726625 kubelet[3245]: I0813 07:22:35.726501 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:36.520019 kubelet[3245]: I0813 07:22:36.520004 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:36.935756 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4273042912.mount: Deactivated successfully. Aug 13 07:22:37.139824 containerd[1920]: time="2025-08-13T07:22:37.139773694Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:37.140046 containerd[1920]: time="2025-08-13T07:22:37.140024131Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.2: active requests=0, bytes read=66352308" Aug 13 07:22:37.140336 containerd[1920]: time="2025-08-13T07:22:37.140298088Z" level=info msg="ImageCreate event name:\"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:37.141495 containerd[1920]: time="2025-08-13T07:22:37.141454204Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:22:37.141979 containerd[1920]: time="2025-08-13T07:22:37.141937335Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" with image id \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\", size \"66352154\" in 2.17095425s" Aug 13 07:22:37.141979 containerd[1920]: time="2025-08-13T07:22:37.141954258Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\"" Aug 13 07:22:37.142934 containerd[1920]: time="2025-08-13T07:22:37.142922022Z" level=info msg="CreateContainer within sandbox \"24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Aug 13 07:22:37.147152 containerd[1920]: time="2025-08-13T07:22:37.147134180Z" level=info msg="CreateContainer within sandbox \"24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"f8435bf743e12918661f5643083a8461b380c0b6deb190c9f2a77361cddfe68a\"" Aug 13 07:22:37.147417 containerd[1920]: time="2025-08-13T07:22:37.147378969Z" level=info msg="StartContainer for \"f8435bf743e12918661f5643083a8461b380c0b6deb190c9f2a77361cddfe68a\"" Aug 13 07:22:37.197905 containerd[1920]: time="2025-08-13T07:22:37.197843225Z" level=info msg="StartContainer for \"f8435bf743e12918661f5643083a8461b380c0b6deb190c9f2a77361cddfe68a\" returns successfully" Aug 13 07:22:38.524712 kubelet[3245]: I0813 07:22:38.524673 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:42.461392 kubelet[3245]: I0813 07:22:42.461299 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:48.342399 containerd[1920]: time="2025-08-13T07:22:48.342289104Z" level=info msg="StopPodSandbox for \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\"" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.361 [WARNING][7456] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0", GenerateName:"calico-kube-controllers-694d7cc5f7-", Namespace:"calico-system", SelfLink:"", UID:"f80e5fda-a5d5-4f6a-9768-261c847bf3ee", ResourceVersion:"1004", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"694d7cc5f7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42", Pod:"calico-kube-controllers-694d7cc5f7-rl69w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.121.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie575646a156", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.361 [INFO][7456] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.361 [INFO][7456] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" iface="eth0" netns="" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.361 [INFO][7456] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.361 [INFO][7456] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.373 [INFO][7475] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.373 [INFO][7475] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.373 [INFO][7475] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.377 [WARNING][7475] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.377 [INFO][7475] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.378 [INFO][7475] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.379957 containerd[1920]: 2025-08-13 07:22:48.379 [INFO][7456] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.380290 containerd[1920]: time="2025-08-13T07:22:48.379983384Z" level=info msg="TearDown network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\" successfully" Aug 13 07:22:48.380290 containerd[1920]: time="2025-08-13T07:22:48.379998048Z" level=info msg="StopPodSandbox for \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\" returns successfully" Aug 13 07:22:48.380335 containerd[1920]: time="2025-08-13T07:22:48.380301847Z" level=info msg="RemovePodSandbox for \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\"" Aug 13 07:22:48.380335 containerd[1920]: time="2025-08-13T07:22:48.380321403Z" level=info msg="Forcibly stopping sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\"" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.399 [WARNING][7498] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0", GenerateName:"calico-kube-controllers-694d7cc5f7-", Namespace:"calico-system", SelfLink:"", UID:"f80e5fda-a5d5-4f6a-9768-261c847bf3ee", ResourceVersion:"1004", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"694d7cc5f7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"da66c0d1733f286eece6a5a2bc2ef26de5c9d27dd0570a52631658645984ab42", Pod:"calico-kube-controllers-694d7cc5f7-rl69w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.121.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie575646a156", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.399 [INFO][7498] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.399 [INFO][7498] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" iface="eth0" netns="" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.399 [INFO][7498] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.400 [INFO][7498] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.411 [INFO][7516] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.411 [INFO][7516] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.411 [INFO][7516] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.416 [WARNING][7516] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.416 [INFO][7516] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" HandleID="k8s-pod-network.1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--kube--controllers--694d7cc5f7--rl69w-eth0" Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.417 [INFO][7516] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.419144 containerd[1920]: 2025-08-13 07:22:48.418 [INFO][7498] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c" Aug 13 07:22:48.419485 containerd[1920]: time="2025-08-13T07:22:48.419141653Z" level=info msg="TearDown network for sandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\" successfully" Aug 13 07:22:48.420981 containerd[1920]: time="2025-08-13T07:22:48.420931381Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:48.420981 containerd[1920]: time="2025-08-13T07:22:48.420961383Z" level=info msg="RemovePodSandbox \"1aa9c30206049db558e8991e86708ebb2fed56150afd93faa6ac3432a327e59c\" returns successfully" Aug 13 07:22:48.421285 containerd[1920]: time="2025-08-13T07:22:48.421239594Z" level=info msg="StopPodSandbox for \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\"" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.439 [WARNING][7538] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"b2d55415-dd30-4e5a-9264-00bc2660bdba", ResourceVersion:"1008", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7", Pod:"calico-apiserver-59fbd88f7c-z2k5j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia774e84ab68", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.440 [INFO][7538] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.440 [INFO][7538] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" iface="eth0" netns="" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.440 [INFO][7538] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.440 [INFO][7538] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.486 [INFO][7551] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.486 [INFO][7551] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.486 [INFO][7551] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.493 [WARNING][7551] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.493 [INFO][7551] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.494 [INFO][7551] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.496894 containerd[1920]: 2025-08-13 07:22:48.495 [INFO][7538] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.497505 containerd[1920]: time="2025-08-13T07:22:48.496929386Z" level=info msg="TearDown network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\" successfully" Aug 13 07:22:48.497505 containerd[1920]: time="2025-08-13T07:22:48.496953980Z" level=info msg="StopPodSandbox for \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\" returns successfully" Aug 13 07:22:48.497505 containerd[1920]: time="2025-08-13T07:22:48.497296933Z" level=info msg="RemovePodSandbox for \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\"" Aug 13 07:22:48.497505 containerd[1920]: time="2025-08-13T07:22:48.497324695Z" level=info msg="Forcibly stopping sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\"" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.527 [WARNING][7579] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"b2d55415-dd30-4e5a-9264-00bc2660bdba", ResourceVersion:"1008", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"558736dca972c0c5c3be3022cb55fb745aa7388c62a76bc8e3429d869748d9a7", Pod:"calico-apiserver-59fbd88f7c-z2k5j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia774e84ab68", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.527 [INFO][7579] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.527 [INFO][7579] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" iface="eth0" netns="" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.527 [INFO][7579] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.527 [INFO][7579] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.546 [INFO][7596] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.546 [INFO][7596] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.546 [INFO][7596] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.553 [WARNING][7596] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.553 [INFO][7596] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" HandleID="k8s-pod-network.47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--z2k5j-eth0" Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.554 [INFO][7596] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.557501 containerd[1920]: 2025-08-13 07:22:48.556 [INFO][7579] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983" Aug 13 07:22:48.558177 containerd[1920]: time="2025-08-13T07:22:48.557503507Z" level=info msg="TearDown network for sandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\" successfully" Aug 13 07:22:48.560949 containerd[1920]: time="2025-08-13T07:22:48.560904488Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:48.560949 containerd[1920]: time="2025-08-13T07:22:48.560944947Z" level=info msg="RemovePodSandbox \"47832eed3f38212d643c1c6a1514e8fad10165f34e9a03ba2ecf4d380d7db983\" returns successfully" Aug 13 07:22:48.561184 containerd[1920]: time="2025-08-13T07:22:48.561142542Z" level=info msg="StopPodSandbox for \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\"" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.578 [WARNING][7625] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6", ResourceVersion:"1026", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765", Pod:"goldmane-58fd7646b9-sjk28", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.121.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicf213437f46", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.578 [INFO][7625] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.578 [INFO][7625] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" iface="eth0" netns="" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.578 [INFO][7625] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.578 [INFO][7625] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.588 [INFO][7642] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.588 [INFO][7642] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.588 [INFO][7642] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.592 [WARNING][7642] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.592 [INFO][7642] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.593 [INFO][7642] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.595016 containerd[1920]: 2025-08-13 07:22:48.594 [INFO][7625] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.595016 containerd[1920]: time="2025-08-13T07:22:48.595004049Z" level=info msg="TearDown network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\" successfully" Aug 13 07:22:48.595016 containerd[1920]: time="2025-08-13T07:22:48.595018579Z" level=info msg="StopPodSandbox for \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\" returns successfully" Aug 13 07:22:48.595546 containerd[1920]: time="2025-08-13T07:22:48.595286196Z" level=info msg="RemovePodSandbox for \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\"" Aug 13 07:22:48.595546 containerd[1920]: time="2025-08-13T07:22:48.595308331Z" level=info msg="Forcibly stopping sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\"" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.613 [WARNING][7665] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"dd85a865-2c7f-4339-b9a7-90c5b37bd6d6", ResourceVersion:"1026", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"24603d3d122c6fa09c46e929e50ee1fe552b801860dbf54a8353989aa0328765", Pod:"goldmane-58fd7646b9-sjk28", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.121.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicf213437f46", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.613 [INFO][7665] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.613 [INFO][7665] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" iface="eth0" netns="" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.613 [INFO][7665] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.613 [INFO][7665] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.624 [INFO][7680] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.624 [INFO][7680] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.624 [INFO][7680] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.629 [WARNING][7680] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.629 [INFO][7680] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" HandleID="k8s-pod-network.f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-goldmane--58fd7646b9--sjk28-eth0" Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.630 [INFO][7680] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.631856 containerd[1920]: 2025-08-13 07:22:48.631 [INFO][7665] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a" Aug 13 07:22:48.632249 containerd[1920]: time="2025-08-13T07:22:48.631892228Z" level=info msg="TearDown network for sandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\" successfully" Aug 13 07:22:48.633517 containerd[1920]: time="2025-08-13T07:22:48.633476085Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:48.633517 containerd[1920]: time="2025-08-13T07:22:48.633508378Z" level=info msg="RemovePodSandbox \"f986c64a633831558b4a4ff6eeb04bca0b8f293170491a2fa2ac2177a88d950a\" returns successfully" Aug 13 07:22:48.633786 containerd[1920]: time="2025-08-13T07:22:48.633747824Z" level=info msg="StopPodSandbox for \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\"" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.651 [WARNING][7705] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"926b03e8-ac14-4763-b7d3-fa81c377f2dd", ResourceVersion:"945", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123", Pod:"coredns-7c65d6cfc9-ggl67", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1b1de5ae811", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.651 [INFO][7705] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.651 [INFO][7705] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" iface="eth0" netns="" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.651 [INFO][7705] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.651 [INFO][7705] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.661 [INFO][7719] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.661 [INFO][7719] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.661 [INFO][7719] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.665 [WARNING][7719] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.665 [INFO][7719] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.666 [INFO][7719] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.668172 containerd[1920]: 2025-08-13 07:22:48.667 [INFO][7705] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.668492 containerd[1920]: time="2025-08-13T07:22:48.668191593Z" level=info msg="TearDown network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\" successfully" Aug 13 07:22:48.668492 containerd[1920]: time="2025-08-13T07:22:48.668207984Z" level=info msg="StopPodSandbox for \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\" returns successfully" Aug 13 07:22:48.668537 containerd[1920]: time="2025-08-13T07:22:48.668505597Z" level=info msg="RemovePodSandbox for \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\"" Aug 13 07:22:48.668537 containerd[1920]: time="2025-08-13T07:22:48.668522039Z" level=info msg="Forcibly stopping sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\"" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.687 [WARNING][7743] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"926b03e8-ac14-4763-b7d3-fa81c377f2dd", ResourceVersion:"945", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"9bc9a9251af9f33f3f3091d9471ddb281fca6e6401a2e00c62a1e30ac7ecc123", Pod:"coredns-7c65d6cfc9-ggl67", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1b1de5ae811", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.687 [INFO][7743] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.687 [INFO][7743] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" iface="eth0" netns="" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.687 [INFO][7743] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.687 [INFO][7743] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.698 [INFO][7758] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.698 [INFO][7758] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.698 [INFO][7758] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.702 [WARNING][7758] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.702 [INFO][7758] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" HandleID="k8s-pod-network.c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--ggl67-eth0" Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.704 [INFO][7758] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.706301 containerd[1920]: 2025-08-13 07:22:48.705 [INFO][7743] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1" Aug 13 07:22:48.706628 containerd[1920]: time="2025-08-13T07:22:48.706326031Z" level=info msg="TearDown network for sandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\" successfully" Aug 13 07:22:48.707826 containerd[1920]: time="2025-08-13T07:22:48.707814440Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:48.707856 containerd[1920]: time="2025-08-13T07:22:48.707843944Z" level=info msg="RemovePodSandbox \"c7f9272860e160ff547f18eabb019da191abdfb7a8e526f627149121e965e2b1\" returns successfully" Aug 13 07:22:48.708046 containerd[1920]: time="2025-08-13T07:22:48.708036202Z" level=info msg="StopPodSandbox for \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\"" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.724 [WARNING][7784] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b90d45be-b93c-47a4-ad3d-74de560838fa", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f", Pod:"csi-node-driver-xnn8d", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.121.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif7fbbc48599", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.724 [INFO][7784] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.724 [INFO][7784] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" iface="eth0" netns="" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.724 [INFO][7784] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.724 [INFO][7784] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.733 [INFO][7802] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.733 [INFO][7802] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.733 [INFO][7802] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.737 [WARNING][7802] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.737 [INFO][7802] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.738 [INFO][7802] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.739568 containerd[1920]: 2025-08-13 07:22:48.738 [INFO][7784] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.739568 containerd[1920]: time="2025-08-13T07:22:48.739566186Z" level=info msg="TearDown network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\" successfully" Aug 13 07:22:48.739923 containerd[1920]: time="2025-08-13T07:22:48.739581272Z" level=info msg="StopPodSandbox for \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\" returns successfully" Aug 13 07:22:48.739923 containerd[1920]: time="2025-08-13T07:22:48.739847562Z" level=info msg="RemovePodSandbox for \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\"" Aug 13 07:22:48.739923 containerd[1920]: time="2025-08-13T07:22:48.739862444Z" level=info msg="Forcibly stopping sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\"" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.757 [WARNING][7825] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b90d45be-b93c-47a4-ad3d-74de560838fa", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"0d7a7a238df2471b79fb277eeab3f66a6a0e3950dde75536d38f0c48b10b5c6f", Pod:"csi-node-driver-xnn8d", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.121.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif7fbbc48599", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.757 [INFO][7825] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.757 [INFO][7825] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" iface="eth0" netns="" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.757 [INFO][7825] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.757 [INFO][7825] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.767 [INFO][7840] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.767 [INFO][7840] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.767 [INFO][7840] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.771 [WARNING][7840] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.771 [INFO][7840] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" HandleID="k8s-pod-network.7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-csi--node--driver--xnn8d-eth0" Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.772 [INFO][7840] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.774028 containerd[1920]: 2025-08-13 07:22:48.773 [INFO][7825] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e" Aug 13 07:22:48.774383 containerd[1920]: time="2025-08-13T07:22:48.774028824Z" level=info msg="TearDown network for sandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\" successfully" Aug 13 07:22:48.775541 containerd[1920]: time="2025-08-13T07:22:48.775527702Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:48.775577 containerd[1920]: time="2025-08-13T07:22:48.775558042Z" level=info msg="RemovePodSandbox \"7f77d4709fb7ee6477c2dd9cad3ee96aabb466ece4f10a29fd941aefd84b115e\" returns successfully" Aug 13 07:22:48.775820 containerd[1920]: time="2025-08-13T07:22:48.775809695Z" level=info msg="StopPodSandbox for \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\"" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.792 [WARNING][7866] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"bb4a7045-3301-41f3-86a2-df653630104c", ResourceVersion:"1011", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b", Pod:"calico-apiserver-59fbd88f7c-qqszn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5327790de75", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.792 [INFO][7866] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.792 [INFO][7866] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" iface="eth0" netns="" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.792 [INFO][7866] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.792 [INFO][7866] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.801 [INFO][7881] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.802 [INFO][7881] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.802 [INFO][7881] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.805 [WARNING][7881] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.805 [INFO][7881] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.806 [INFO][7881] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.807675 containerd[1920]: 2025-08-13 07:22:48.806 [INFO][7866] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.808025 containerd[1920]: time="2025-08-13T07:22:48.807703821Z" level=info msg="TearDown network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\" successfully" Aug 13 07:22:48.808025 containerd[1920]: time="2025-08-13T07:22:48.807718613Z" level=info msg="StopPodSandbox for \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\" returns successfully" Aug 13 07:22:48.808025 containerd[1920]: time="2025-08-13T07:22:48.807972380Z" level=info msg="RemovePodSandbox for \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\"" Aug 13 07:22:48.808025 containerd[1920]: time="2025-08-13T07:22:48.807987372Z" level=info msg="Forcibly stopping sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\"" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.825 [WARNING][7905] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0", GenerateName:"calico-apiserver-59fbd88f7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"bb4a7045-3301-41f3-86a2-df653630104c", ResourceVersion:"1011", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 22, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"59fbd88f7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"abbe59f7415f1a9e759171ed0095956c72ef05fc65c8a73c12db9cdddc6ef56b", Pod:"calico-apiserver-59fbd88f7c-qqszn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.121.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5327790de75", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.825 [INFO][7905] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.825 [INFO][7905] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" iface="eth0" netns="" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.825 [INFO][7905] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.825 [INFO][7905] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.835 [INFO][7922] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.835 [INFO][7922] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.835 [INFO][7922] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.840 [WARNING][7922] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.840 [INFO][7922] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" HandleID="k8s-pod-network.a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-calico--apiserver--59fbd88f7c--qqszn-eth0" Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.841 [INFO][7922] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.842812 containerd[1920]: 2025-08-13 07:22:48.841 [INFO][7905] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451" Aug 13 07:22:48.842812 containerd[1920]: time="2025-08-13T07:22:48.842801631Z" level=info msg="TearDown network for sandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\" successfully" Aug 13 07:22:48.844247 containerd[1920]: time="2025-08-13T07:22:48.844234320Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:48.844273 containerd[1920]: time="2025-08-13T07:22:48.844264924Z" level=info msg="RemovePodSandbox \"a86ea392232da98cbc5682188d31c7e7c26058923c82c89dc1518cff02dfd451\" returns successfully" Aug 13 07:22:48.844547 containerd[1920]: time="2025-08-13T07:22:48.844537269Z" level=info msg="StopPodSandbox for \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\"" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.861 [WARNING][7947] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"89af5e87-c375-4eee-977b-cf953f1e3e5b", ResourceVersion:"908", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b", Pod:"coredns-7c65d6cfc9-brr5j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6514525fc53", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.861 [INFO][7947] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.861 [INFO][7947] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" iface="eth0" netns="" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.861 [INFO][7947] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.861 [INFO][7947] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.871 [INFO][7965] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.871 [INFO][7965] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.871 [INFO][7965] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.875 [WARNING][7965] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.875 [INFO][7965] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.876 [INFO][7965] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.878024 containerd[1920]: 2025-08-13 07:22:48.877 [INFO][7947] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.878024 containerd[1920]: time="2025-08-13T07:22:48.878011960Z" level=info msg="TearDown network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\" successfully" Aug 13 07:22:48.878024 containerd[1920]: time="2025-08-13T07:22:48.878026514Z" level=info msg="StopPodSandbox for \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\" returns successfully" Aug 13 07:22:48.878460 containerd[1920]: time="2025-08-13T07:22:48.878336163Z" level=info msg="RemovePodSandbox for \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\"" Aug 13 07:22:48.878460 containerd[1920]: time="2025-08-13T07:22:48.878355191Z" level=info msg="Forcibly stopping sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\"" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.895 [WARNING][7988] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"89af5e87-c375-4eee-977b-cf953f1e3e5b", ResourceVersion:"908", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 21, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-3ca3f1276f", ContainerID:"87114c05ccbe40b870470e107b5e9b8d4f0f03877b8de614a4f9b46e3f97da9b", Pod:"coredns-7c65d6cfc9-brr5j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.121.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6514525fc53", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.895 [INFO][7988] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.895 [INFO][7988] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" iface="eth0" netns="" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.895 [INFO][7988] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.895 [INFO][7988] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.905 [INFO][8004] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.905 [INFO][8004] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.905 [INFO][8004] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.910 [WARNING][8004] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.910 [INFO][8004] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" HandleID="k8s-pod-network.74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-coredns--7c65d6cfc9--brr5j-eth0" Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.911 [INFO][8004] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:48.913192 containerd[1920]: 2025-08-13 07:22:48.912 [INFO][7988] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296" Aug 13 07:22:48.913192 containerd[1920]: time="2025-08-13T07:22:48.913185393Z" level=info msg="TearDown network for sandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\" successfully" Aug 13 07:22:48.914601 containerd[1920]: time="2025-08-13T07:22:48.914589547Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:48.914633 containerd[1920]: time="2025-08-13T07:22:48.914618711Z" level=info msg="RemovePodSandbox \"74e2491ed4e5e3766ab192400717c4a71cd4680ae511be3c515d8db8dd3e3296\" returns successfully" Aug 13 07:22:48.914867 containerd[1920]: time="2025-08-13T07:22:48.914856359Z" level=info msg="StopPodSandbox for \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\"" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.938 [WARNING][8028] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.938 [INFO][8028] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.938 [INFO][8028] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" iface="eth0" netns="" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.938 [INFO][8028] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.939 [INFO][8028] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.993 [INFO][8043] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.993 [INFO][8043] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:48.993 [INFO][8043] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:49.010 [WARNING][8043] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:49.010 [INFO][8043] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:49.013 [INFO][8043] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:49.020800 containerd[1920]: 2025-08-13 07:22:49.017 [INFO][8028] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.022107 containerd[1920]: time="2025-08-13T07:22:49.020849936Z" level=info msg="TearDown network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\" successfully" Aug 13 07:22:49.022107 containerd[1920]: time="2025-08-13T07:22:49.020908765Z" level=info msg="StopPodSandbox for \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\" returns successfully" Aug 13 07:22:49.022107 containerd[1920]: time="2025-08-13T07:22:49.021890814Z" level=info msg="RemovePodSandbox for \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\"" Aug 13 07:22:49.022107 containerd[1920]: time="2025-08-13T07:22:49.021977822Z" level=info msg="Forcibly stopping sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\"" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.086 [WARNING][8070] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" WorkloadEndpoint="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.087 [INFO][8070] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.087 [INFO][8070] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" iface="eth0" netns="" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.087 [INFO][8070] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.087 [INFO][8070] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.104 [INFO][8088] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.104 [INFO][8088] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.104 [INFO][8088] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.108 [WARNING][8088] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.108 [INFO][8088] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" HandleID="k8s-pod-network.47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Workload="ci--4081.3.5--a--3ca3f1276f-k8s-whisker--5f8549b86f--sqvnt-eth0" Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.110 [INFO][8088] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:22:49.111722 containerd[1920]: 2025-08-13 07:22:49.110 [INFO][8070] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56" Aug 13 07:22:49.112013 containerd[1920]: time="2025-08-13T07:22:49.111753426Z" level=info msg="TearDown network for sandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\" successfully" Aug 13 07:22:49.114387 containerd[1920]: time="2025-08-13T07:22:49.114371739Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:22:49.114425 containerd[1920]: time="2025-08-13T07:22:49.114401049Z" level=info msg="RemovePodSandbox \"47787386954f35448db44abee710d193060aabac48aac47fefbd67dadb3ecd56\" returns successfully" Aug 13 07:22:56.009875 kubelet[3245]: I0813 07:22:56.009749 3245 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:22:56.040783 kubelet[3245]: I0813 07:22:56.040750 3245 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-58fd7646b9-sjk28" podStartSLOduration=44.499231532 podStartE2EDuration="52.040739109s" podCreationTimestamp="2025-08-13 07:22:04 +0000 UTC" firstStartedPulling="2025-08-13 07:22:29.600879169 +0000 UTC m=+41.323841054" lastFinishedPulling="2025-08-13 07:22:37.142386745 +0000 UTC m=+48.865348631" observedRunningTime="2025-08-13 07:22:37.536736476 +0000 UTC m=+49.259698362" watchObservedRunningTime="2025-08-13 07:22:56.040739109 +0000 UTC m=+67.763700992" Aug 13 07:23:55.113490 update_engine[1913]: I20250813 07:23:55.113353 1913 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Aug 13 07:23:55.113490 update_engine[1913]: I20250813 07:23:55.113458 1913 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Aug 13 07:23:55.114926 update_engine[1913]: I20250813 07:23:55.113847 1913 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Aug 13 07:23:55.114926 update_engine[1913]: I20250813 07:23:55.114892 1913 omaha_request_params.cc:62] Current group set to lts Aug 13 07:23:55.115280 update_engine[1913]: I20250813 07:23:55.115209 1913 update_attempter.cc:499] Already updated boot flags. Skipping. Aug 13 07:23:55.115280 update_engine[1913]: I20250813 07:23:55.115245 1913 update_attempter.cc:643] Scheduling an action processor start. Aug 13 07:23:55.115572 update_engine[1913]: I20250813 07:23:55.115283 1913 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Aug 13 07:23:55.115572 update_engine[1913]: I20250813 07:23:55.115356 1913 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Aug 13 07:23:55.115572 update_engine[1913]: I20250813 07:23:55.115525 1913 omaha_request_action.cc:271] Posting an Omaha request to disabled Aug 13 07:23:55.115572 update_engine[1913]: I20250813 07:23:55.115556 1913 omaha_request_action.cc:272] Request: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.115572 update_engine[1913]: Aug 13 07:23:55.116588 update_engine[1913]: I20250813 07:23:55.115574 1913 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 07:23:55.116692 locksmithd[1951]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Aug 13 07:23:55.118037 update_engine[1913]: I20250813 07:23:55.118026 1913 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 07:23:55.118289 update_engine[1913]: I20250813 07:23:55.118249 1913 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 07:23:55.119228 update_engine[1913]: E20250813 07:23:55.119213 1913 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 07:23:55.119260 update_engine[1913]: I20250813 07:23:55.119244 1913 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Aug 13 07:24:05.112303 update_engine[1913]: I20250813 07:24:05.112237 1913 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 07:24:05.114372 update_engine[1913]: I20250813 07:24:05.112370 1913 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 07:24:05.114372 update_engine[1913]: I20250813 07:24:05.112517 1913 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 07:24:05.114372 update_engine[1913]: E20250813 07:24:05.113059 1913 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 07:24:05.114372 update_engine[1913]: I20250813 07:24:05.113101 1913 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Aug 13 07:24:15.112438 update_engine[1913]: I20250813 07:24:15.112185 1913 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 07:24:15.113501 update_engine[1913]: I20250813 07:24:15.112727 1913 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 07:24:15.113501 update_engine[1913]: I20250813 07:24:15.113262 1913 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 07:24:15.114164 update_engine[1913]: E20250813 07:24:15.114027 1913 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 07:24:15.114349 update_engine[1913]: I20250813 07:24:15.114202 1913 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Aug 13 07:24:25.112457 update_engine[1913]: I20250813 07:24:25.112293 1913 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 07:24:25.113536 update_engine[1913]: I20250813 07:24:25.112839 1913 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 07:24:25.113536 update_engine[1913]: I20250813 07:24:25.113387 1913 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 07:24:25.114252 update_engine[1913]: E20250813 07:24:25.114137 1913 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 07:24:25.114464 update_engine[1913]: I20250813 07:24:25.114276 1913 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Aug 13 07:24:25.114464 update_engine[1913]: I20250813 07:24:25.114306 1913 omaha_request_action.cc:617] Omaha request response: Aug 13 07:24:25.114754 update_engine[1913]: E20250813 07:24:25.114466 1913 omaha_request_action.cc:636] Omaha request network transfer failed. Aug 13 07:24:25.114754 update_engine[1913]: I20250813 07:24:25.114517 1913 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Aug 13 07:24:25.114754 update_engine[1913]: I20250813 07:24:25.114537 1913 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Aug 13 07:24:25.114754 update_engine[1913]: I20250813 07:24:25.114552 1913 update_attempter.cc:306] Processing Done. Aug 13 07:24:25.114754 update_engine[1913]: E20250813 07:24:25.114583 1913 update_attempter.cc:619] Update failed. Aug 13 07:24:25.114754 update_engine[1913]: I20250813 07:24:25.114600 1913 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Aug 13 07:24:25.114754 update_engine[1913]: I20250813 07:24:25.114616 1913 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Aug 13 07:24:25.114754 update_engine[1913]: I20250813 07:24:25.114632 1913 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Aug 13 07:24:25.115500 update_engine[1913]: I20250813 07:24:25.114787 1913 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Aug 13 07:24:25.115500 update_engine[1913]: I20250813 07:24:25.114853 1913 omaha_request_action.cc:271] Posting an Omaha request to disabled Aug 13 07:24:25.115500 update_engine[1913]: I20250813 07:24:25.114873 1913 omaha_request_action.cc:272] Request: Aug 13 07:24:25.115500 update_engine[1913]: Aug 13 07:24:25.115500 update_engine[1913]: Aug 13 07:24:25.115500 update_engine[1913]: Aug 13 07:24:25.115500 update_engine[1913]: Aug 13 07:24:25.115500 update_engine[1913]: Aug 13 07:24:25.115500 update_engine[1913]: Aug 13 07:24:25.115500 update_engine[1913]: I20250813 07:24:25.114890 1913 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 07:24:25.115500 update_engine[1913]: I20250813 07:24:25.115334 1913 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 07:24:25.116411 update_engine[1913]: I20250813 07:24:25.115738 1913 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 07:24:25.116517 locksmithd[1951]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Aug 13 07:24:25.117261 update_engine[1913]: E20250813 07:24:25.116496 1913 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 07:24:25.117261 update_engine[1913]: I20250813 07:24:25.116629 1913 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Aug 13 07:24:25.117261 update_engine[1913]: I20250813 07:24:25.116658 1913 omaha_request_action.cc:617] Omaha request response: Aug 13 07:24:25.117261 update_engine[1913]: I20250813 07:24:25.116678 1913 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Aug 13 07:24:25.117261 update_engine[1913]: I20250813 07:24:25.116693 1913 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Aug 13 07:24:25.117261 update_engine[1913]: I20250813 07:24:25.116710 1913 update_attempter.cc:306] Processing Done. Aug 13 07:24:25.117261 update_engine[1913]: I20250813 07:24:25.116727 1913 update_attempter.cc:310] Error event sent. Aug 13 07:24:25.117261 update_engine[1913]: I20250813 07:24:25.116760 1913 update_check_scheduler.cc:74] Next update check in 48m12s Aug 13 07:24:25.117959 locksmithd[1951]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Aug 13 07:28:00.308768 systemd[1]: Started sshd@9-145.40.90.243:22-147.75.109.163:44330.service - OpenSSH per-connection server daemon (147.75.109.163:44330). Aug 13 07:28:00.349380 sshd[9327]: Accepted publickey for core from 147.75.109.163 port 44330 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:00.351342 sshd[9327]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:00.358249 systemd-logind[1908]: New session 12 of user core. Aug 13 07:28:00.373950 systemd[1]: Started session-12.scope - Session 12 of User core. Aug 13 07:28:00.540428 sshd[9327]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:00.548520 systemd[1]: sshd@9-145.40.90.243:22-147.75.109.163:44330.service: Deactivated successfully. Aug 13 07:28:00.551040 systemd-logind[1908]: Session 12 logged out. Waiting for processes to exit. Aug 13 07:28:00.551240 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 07:28:00.552221 systemd-logind[1908]: Removed session 12. Aug 13 07:28:05.563336 systemd[1]: Started sshd@10-145.40.90.243:22-147.75.109.163:44344.service - OpenSSH per-connection server daemon (147.75.109.163:44344). Aug 13 07:28:05.599077 sshd[9411]: Accepted publickey for core from 147.75.109.163 port 44344 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:05.602326 sshd[9411]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:05.604824 systemd-logind[1908]: New session 13 of user core. Aug 13 07:28:05.618318 systemd[1]: Started session-13.scope - Session 13 of User core. Aug 13 07:28:05.697379 sshd[9411]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:05.698931 systemd[1]: sshd@10-145.40.90.243:22-147.75.109.163:44344.service: Deactivated successfully. Aug 13 07:28:05.700280 systemd-logind[1908]: Session 13 logged out. Waiting for processes to exit. Aug 13 07:28:05.700372 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 07:28:05.700963 systemd-logind[1908]: Removed session 13. Aug 13 07:28:10.717781 systemd[1]: Started sshd@11-145.40.90.243:22-147.75.109.163:49726.service - OpenSSH per-connection server daemon (147.75.109.163:49726). Aug 13 07:28:10.750382 sshd[9441]: Accepted publickey for core from 147.75.109.163 port 49726 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:10.751150 sshd[9441]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:10.753957 systemd-logind[1908]: New session 14 of user core. Aug 13 07:28:10.765266 systemd[1]: Started session-14.scope - Session 14 of User core. Aug 13 07:28:10.851934 sshd[9441]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:10.867803 systemd[1]: Started sshd@12-145.40.90.243:22-147.75.109.163:49740.service - OpenSSH per-connection server daemon (147.75.109.163:49740). Aug 13 07:28:10.869497 systemd[1]: sshd@11-145.40.90.243:22-147.75.109.163:49726.service: Deactivated successfully. Aug 13 07:28:10.873538 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 07:28:10.876991 systemd-logind[1908]: Session 14 logged out. Waiting for processes to exit. Aug 13 07:28:10.879929 systemd-logind[1908]: Removed session 14. Aug 13 07:28:10.939349 sshd[9467]: Accepted publickey for core from 147.75.109.163 port 49740 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:10.942779 sshd[9467]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:10.954866 systemd-logind[1908]: New session 15 of user core. Aug 13 07:28:10.970784 systemd[1]: Started session-15.scope - Session 15 of User core. Aug 13 07:28:11.111795 sshd[9467]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:11.124309 systemd[1]: Started sshd@13-145.40.90.243:22-147.75.109.163:49754.service - OpenSSH per-connection server daemon (147.75.109.163:49754). Aug 13 07:28:11.124648 systemd[1]: sshd@12-145.40.90.243:22-147.75.109.163:49740.service: Deactivated successfully. Aug 13 07:28:11.125537 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 07:28:11.126220 systemd-logind[1908]: Session 15 logged out. Waiting for processes to exit. Aug 13 07:28:11.126826 systemd-logind[1908]: Removed session 15. Aug 13 07:28:11.153922 sshd[9493]: Accepted publickey for core from 147.75.109.163 port 49754 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:11.157512 sshd[9493]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:11.169691 systemd-logind[1908]: New session 16 of user core. Aug 13 07:28:11.187731 systemd[1]: Started session-16.scope - Session 16 of User core. Aug 13 07:28:11.339594 sshd[9493]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:11.341558 systemd[1]: sshd@13-145.40.90.243:22-147.75.109.163:49754.service: Deactivated successfully. Aug 13 07:28:11.343443 systemd-logind[1908]: Session 16 logged out. Waiting for processes to exit. Aug 13 07:28:11.343522 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 07:28:11.344302 systemd-logind[1908]: Removed session 16. Aug 13 07:28:16.361674 systemd[1]: Started sshd@14-145.40.90.243:22-147.75.109.163:49756.service - OpenSSH per-connection server daemon (147.75.109.163:49756). Aug 13 07:28:16.417227 sshd[9557]: Accepted publickey for core from 147.75.109.163 port 49756 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:16.417884 sshd[9557]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:16.420590 systemd-logind[1908]: New session 17 of user core. Aug 13 07:28:16.435813 systemd[1]: Started session-17.scope - Session 17 of User core. Aug 13 07:28:16.582984 sshd[9557]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:16.584594 systemd[1]: sshd@14-145.40.90.243:22-147.75.109.163:49756.service: Deactivated successfully. Aug 13 07:28:16.586025 systemd-logind[1908]: Session 17 logged out. Waiting for processes to exit. Aug 13 07:28:16.586161 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 07:28:16.586756 systemd-logind[1908]: Removed session 17. Aug 13 07:28:21.599344 systemd[1]: Started sshd@15-145.40.90.243:22-147.75.109.163:36366.service - OpenSSH per-connection server daemon (147.75.109.163:36366). Aug 13 07:28:21.628127 sshd[9587]: Accepted publickey for core from 147.75.109.163 port 36366 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:21.629020 sshd[9587]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:21.632050 systemd-logind[1908]: New session 18 of user core. Aug 13 07:28:21.648366 systemd[1]: Started session-18.scope - Session 18 of User core. Aug 13 07:28:21.737754 sshd[9587]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:21.739279 systemd[1]: sshd@15-145.40.90.243:22-147.75.109.163:36366.service: Deactivated successfully. Aug 13 07:28:21.740796 systemd-logind[1908]: Session 18 logged out. Waiting for processes to exit. Aug 13 07:28:21.740890 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 07:28:21.741629 systemd-logind[1908]: Removed session 18. Aug 13 07:28:26.763373 systemd[1]: Started sshd@16-145.40.90.243:22-147.75.109.163:36372.service - OpenSSH per-connection server daemon (147.75.109.163:36372). Aug 13 07:28:26.793431 sshd[9618]: Accepted publickey for core from 147.75.109.163 port 36372 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:26.794608 sshd[9618]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:26.798671 systemd-logind[1908]: New session 19 of user core. Aug 13 07:28:26.814782 systemd[1]: Started session-19.scope - Session 19 of User core. Aug 13 07:28:26.952479 sshd[9618]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:26.953983 systemd[1]: sshd@16-145.40.90.243:22-147.75.109.163:36372.service: Deactivated successfully. Aug 13 07:28:26.955389 systemd-logind[1908]: Session 19 logged out. Waiting for processes to exit. Aug 13 07:28:26.955464 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 07:28:26.955993 systemd-logind[1908]: Removed session 19. Aug 13 07:28:31.973259 systemd[1]: Started sshd@17-145.40.90.243:22-147.75.109.163:40172.service - OpenSSH per-connection server daemon (147.75.109.163:40172). Aug 13 07:28:32.002342 sshd[9676]: Accepted publickey for core from 147.75.109.163 port 40172 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:32.003261 sshd[9676]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:32.006302 systemd-logind[1908]: New session 20 of user core. Aug 13 07:28:32.020356 systemd[1]: Started session-20.scope - Session 20 of User core. Aug 13 07:28:32.107823 sshd[9676]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:32.118847 systemd[1]: Started sshd@18-145.40.90.243:22-147.75.109.163:40188.service - OpenSSH per-connection server daemon (147.75.109.163:40188). Aug 13 07:28:32.120200 systemd[1]: sshd@17-145.40.90.243:22-147.75.109.163:40172.service: Deactivated successfully. Aug 13 07:28:32.125367 systemd-logind[1908]: Session 20 logged out. Waiting for processes to exit. Aug 13 07:28:32.126901 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 07:28:32.129991 systemd-logind[1908]: Removed session 20. Aug 13 07:28:32.191059 sshd[9700]: Accepted publickey for core from 147.75.109.163 port 40188 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:32.194524 sshd[9700]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:32.206378 systemd-logind[1908]: New session 21 of user core. Aug 13 07:28:32.231814 systemd[1]: Started session-21.scope - Session 21 of User core. Aug 13 07:28:32.381363 sshd[9700]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:32.395841 systemd[1]: Started sshd@19-145.40.90.243:22-147.75.109.163:40192.service - OpenSSH per-connection server daemon (147.75.109.163:40192). Aug 13 07:28:32.397495 systemd[1]: sshd@18-145.40.90.243:22-147.75.109.163:40188.service: Deactivated successfully. Aug 13 07:28:32.401572 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 07:28:32.404923 systemd-logind[1908]: Session 21 logged out. Waiting for processes to exit. Aug 13 07:28:32.407859 systemd-logind[1908]: Removed session 21. Aug 13 07:28:32.467392 sshd[9726]: Accepted publickey for core from 147.75.109.163 port 40192 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:32.470899 sshd[9726]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:32.483822 systemd-logind[1908]: New session 22 of user core. Aug 13 07:28:32.502837 systemd[1]: Started session-22.scope - Session 22 of User core. Aug 13 07:28:33.584482 sshd[9726]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:33.598297 systemd[1]: Started sshd@20-145.40.90.243:22-147.75.109.163:40206.service - OpenSSH per-connection server daemon (147.75.109.163:40206). Aug 13 07:28:33.598675 systemd[1]: sshd@19-145.40.90.243:22-147.75.109.163:40192.service: Deactivated successfully. Aug 13 07:28:33.599597 systemd[1]: session-22.scope: Deactivated successfully. Aug 13 07:28:33.600303 systemd-logind[1908]: Session 22 logged out. Waiting for processes to exit. Aug 13 07:28:33.600982 systemd-logind[1908]: Removed session 22. Aug 13 07:28:33.627807 sshd[9787]: Accepted publickey for core from 147.75.109.163 port 40206 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:33.631352 sshd[9787]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:33.643689 systemd-logind[1908]: New session 23 of user core. Aug 13 07:28:33.665846 systemd[1]: Started session-23.scope - Session 23 of User core. Aug 13 07:28:33.862173 sshd[9787]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:33.875957 systemd[1]: Started sshd@21-145.40.90.243:22-147.75.109.163:40212.service - OpenSSH per-connection server daemon (147.75.109.163:40212). Aug 13 07:28:33.877511 systemd[1]: sshd@20-145.40.90.243:22-147.75.109.163:40206.service: Deactivated successfully. Aug 13 07:28:33.882175 systemd[1]: session-23.scope: Deactivated successfully. Aug 13 07:28:33.883842 systemd-logind[1908]: Session 23 logged out. Waiting for processes to exit. Aug 13 07:28:33.885909 systemd-logind[1908]: Removed session 23. Aug 13 07:28:33.924403 sshd[9814]: Accepted publickey for core from 147.75.109.163 port 40212 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:33.928239 sshd[9814]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:33.940519 systemd-logind[1908]: New session 24 of user core. Aug 13 07:28:33.956913 systemd[1]: Started session-24.scope - Session 24 of User core. Aug 13 07:28:34.105938 sshd[9814]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:34.107675 systemd[1]: sshd@21-145.40.90.243:22-147.75.109.163:40212.service: Deactivated successfully. Aug 13 07:28:34.109029 systemd-logind[1908]: Session 24 logged out. Waiting for processes to exit. Aug 13 07:28:34.109166 systemd[1]: session-24.scope: Deactivated successfully. Aug 13 07:28:34.109745 systemd-logind[1908]: Removed session 24. Aug 13 07:28:39.126355 systemd[1]: Started sshd@22-145.40.90.243:22-147.75.109.163:52624.service - OpenSSH per-connection server daemon (147.75.109.163:52624). Aug 13 07:28:39.155806 sshd[9885]: Accepted publickey for core from 147.75.109.163 port 52624 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:39.156822 sshd[9885]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:39.160777 systemd-logind[1908]: New session 25 of user core. Aug 13 07:28:39.170414 systemd[1]: Started session-25.scope - Session 25 of User core. Aug 13 07:28:39.260810 sshd[9885]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:39.262228 systemd[1]: sshd@22-145.40.90.243:22-147.75.109.163:52624.service: Deactivated successfully. Aug 13 07:28:39.263703 systemd-logind[1908]: Session 25 logged out. Waiting for processes to exit. Aug 13 07:28:39.263829 systemd[1]: session-25.scope: Deactivated successfully. Aug 13 07:28:39.264553 systemd-logind[1908]: Removed session 25. Aug 13 07:28:44.277277 systemd[1]: Started sshd@23-145.40.90.243:22-147.75.109.163:52640.service - OpenSSH per-connection server daemon (147.75.109.163:52640). Aug 13 07:28:44.306556 sshd[9944]: Accepted publickey for core from 147.75.109.163 port 52640 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:44.307431 sshd[9944]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:44.310726 systemd-logind[1908]: New session 26 of user core. Aug 13 07:28:44.328449 systemd[1]: Started session-26.scope - Session 26 of User core. Aug 13 07:28:44.458095 sshd[9944]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:44.459531 systemd[1]: sshd@23-145.40.90.243:22-147.75.109.163:52640.service: Deactivated successfully. Aug 13 07:28:44.460926 systemd-logind[1908]: Session 26 logged out. Waiting for processes to exit. Aug 13 07:28:44.461017 systemd[1]: session-26.scope: Deactivated successfully. Aug 13 07:28:44.461576 systemd-logind[1908]: Removed session 26. Aug 13 07:28:49.482364 systemd[1]: Started sshd@24-145.40.90.243:22-147.75.109.163:54048.service - OpenSSH per-connection server daemon (147.75.109.163:54048). Aug 13 07:28:49.512620 sshd[9973]: Accepted publickey for core from 147.75.109.163 port 54048 ssh2: RSA SHA256:J9bO4QOv3eMMxMAPUK7J8OKu4RrTNshNa8HZDHJxfKY Aug 13 07:28:49.513808 sshd[9973]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:28:49.518540 systemd-logind[1908]: New session 27 of user core. Aug 13 07:28:49.539765 systemd[1]: Started session-27.scope - Session 27 of User core. Aug 13 07:28:49.625451 sshd[9973]: pam_unix(sshd:session): session closed for user core Aug 13 07:28:49.627142 systemd[1]: sshd@24-145.40.90.243:22-147.75.109.163:54048.service: Deactivated successfully. Aug 13 07:28:49.628692 systemd-logind[1908]: Session 27 logged out. Waiting for processes to exit. Aug 13 07:28:49.628779 systemd[1]: session-27.scope: Deactivated successfully. Aug 13 07:28:49.629424 systemd-logind[1908]: Removed session 27.