Sep 13 01:20:31.026584 kernel: Linux version 6.6.106-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Sep 12 22:30:50 -00 2025 Sep 13 01:20:31.026598 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:20:31.026605 kernel: BIOS-provided physical RAM map: Sep 13 01:20:31.026609 kernel: BIOS-e820: [mem 0x0000000000000000-0x00000000000997ff] usable Sep 13 01:20:31.026613 kernel: BIOS-e820: [mem 0x0000000000099800-0x000000000009ffff] reserved Sep 13 01:20:31.026617 kernel: BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved Sep 13 01:20:31.026622 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003fffffff] usable Sep 13 01:20:31.026626 kernel: BIOS-e820: [mem 0x0000000040000000-0x00000000403fffff] reserved Sep 13 01:20:31.026630 kernel: BIOS-e820: [mem 0x0000000040400000-0x00000000825bdfff] usable Sep 13 01:20:31.026635 kernel: BIOS-e820: [mem 0x00000000825be000-0x00000000825befff] ACPI NVS Sep 13 01:20:31.026639 kernel: BIOS-e820: [mem 0x00000000825bf000-0x00000000825bffff] reserved Sep 13 01:20:31.026644 kernel: BIOS-e820: [mem 0x00000000825c0000-0x000000008afcdfff] usable Sep 13 01:20:31.026648 kernel: BIOS-e820: [mem 0x000000008afce000-0x000000008c0b2fff] reserved Sep 13 01:20:31.026653 kernel: BIOS-e820: [mem 0x000000008c0b3000-0x000000008c23bfff] usable Sep 13 01:20:31.026658 kernel: BIOS-e820: [mem 0x000000008c23c000-0x000000008c66dfff] ACPI NVS Sep 13 01:20:31.026663 kernel: BIOS-e820: [mem 0x000000008c66e000-0x000000008eefefff] reserved Sep 13 01:20:31.026668 kernel: BIOS-e820: [mem 0x000000008eeff000-0x000000008eefffff] usable Sep 13 01:20:31.026673 kernel: BIOS-e820: [mem 0x000000008ef00000-0x000000008fffffff] reserved Sep 13 01:20:31.026678 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 13 01:20:31.026682 kernel: BIOS-e820: [mem 0x00000000fe000000-0x00000000fe010fff] reserved Sep 13 01:20:31.026687 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec00fff] reserved Sep 13 01:20:31.026692 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Sep 13 01:20:31.026696 kernel: BIOS-e820: [mem 0x00000000ff000000-0x00000000ffffffff] reserved Sep 13 01:20:31.026701 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000086effffff] usable Sep 13 01:20:31.026706 kernel: NX (Execute Disable) protection: active Sep 13 01:20:31.026711 kernel: APIC: Static calls initialized Sep 13 01:20:31.026715 kernel: SMBIOS 3.2.1 present. Sep 13 01:20:31.026720 kernel: DMI: Supermicro SYS-5019C-MR-PH004/X11SCM-F, BIOS 2.6 12/03/2024 Sep 13 01:20:31.026726 kernel: tsc: Detected 3400.000 MHz processor Sep 13 01:20:31.026731 kernel: tsc: Detected 3399.906 MHz TSC Sep 13 01:20:31.026736 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 13 01:20:31.026741 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 13 01:20:31.026746 kernel: last_pfn = 0x86f000 max_arch_pfn = 0x400000000 Sep 13 01:20:31.026751 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 23), built from 10 variable MTRRs Sep 13 01:20:31.026756 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 13 01:20:31.026760 kernel: last_pfn = 0x8ef00 max_arch_pfn = 0x400000000 Sep 13 01:20:31.026765 kernel: Using GB pages for direct mapping Sep 13 01:20:31.026771 kernel: ACPI: Early table checksum verification disabled Sep 13 01:20:31.026776 kernel: ACPI: RSDP 0x00000000000F05B0 000024 (v02 SUPERM) Sep 13 01:20:31.026781 kernel: ACPI: XSDT 0x000000008C54F0C8 00010C (v01 SUPERM SUPERM 01072009 AMI 00010013) Sep 13 01:20:31.026788 kernel: ACPI: FACP 0x000000008C58B670 000114 (v06 01072009 AMI 00010013) Sep 13 01:20:31.026793 kernel: ACPI: DSDT 0x000000008C54F268 03C404 (v02 SUPERM SMCI--MB 01072009 INTL 20160527) Sep 13 01:20:31.026798 kernel: ACPI: FACS 0x000000008C66DF80 000040 Sep 13 01:20:31.026803 kernel: ACPI: APIC 0x000000008C58B788 00012C (v04 01072009 AMI 00010013) Sep 13 01:20:31.026810 kernel: ACPI: FPDT 0x000000008C58B8B8 000044 (v01 01072009 AMI 00010013) Sep 13 01:20:31.026815 kernel: ACPI: FIDT 0x000000008C58B900 00009C (v01 SUPERM SMCI--MB 01072009 AMI 00010013) Sep 13 01:20:31.026820 kernel: ACPI: MCFG 0x000000008C58B9A0 00003C (v01 SUPERM SMCI--MB 01072009 MSFT 00000097) Sep 13 01:20:31.026825 kernel: ACPI: SPMI 0x000000008C58B9E0 000041 (v05 SUPERM SMCI--MB 00000000 AMI. 00000000) Sep 13 01:20:31.026830 kernel: ACPI: SSDT 0x000000008C58BA28 001B1C (v02 CpuRef CpuSsdt 00003000 INTL 20160527) Sep 13 01:20:31.026835 kernel: ACPI: SSDT 0x000000008C58D548 0031C6 (v02 SaSsdt SaSsdt 00003000 INTL 20160527) Sep 13 01:20:31.026840 kernel: ACPI: SSDT 0x000000008C590710 00232B (v02 PegSsd PegSsdt 00001000 INTL 20160527) Sep 13 01:20:31.026846 kernel: ACPI: HPET 0x000000008C592A40 000038 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026852 kernel: ACPI: SSDT 0x000000008C592A78 000FAE (v02 SUPERM Ther_Rvp 00001000 INTL 20160527) Sep 13 01:20:31.026857 kernel: ACPI: SSDT 0x000000008C593A28 0008F4 (v02 INTEL xh_mossb 00000000 INTL 20160527) Sep 13 01:20:31.026862 kernel: ACPI: UEFI 0x000000008C594320 000042 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026867 kernel: ACPI: LPIT 0x000000008C594368 000094 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026872 kernel: ACPI: SSDT 0x000000008C594400 0027DE (v02 SUPERM PtidDevc 00001000 INTL 20160527) Sep 13 01:20:31.026877 kernel: ACPI: SSDT 0x000000008C596BE0 0014E2 (v02 SUPERM TbtTypeC 00000000 INTL 20160527) Sep 13 01:20:31.026882 kernel: ACPI: DBGP 0x000000008C5980C8 000034 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026887 kernel: ACPI: DBG2 0x000000008C598100 000054 (v00 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026893 kernel: ACPI: SSDT 0x000000008C598158 001B67 (v02 SUPERM UsbCTabl 00001000 INTL 20160527) Sep 13 01:20:31.026899 kernel: ACPI: DMAR 0x000000008C599CC0 000070 (v01 INTEL EDK2 00000002 01000013) Sep 13 01:20:31.026904 kernel: ACPI: SSDT 0x000000008C599D30 000144 (v02 Intel ADebTabl 00001000 INTL 20160527) Sep 13 01:20:31.026909 kernel: ACPI: TPM2 0x000000008C599E78 000034 (v04 SUPERM SMCI--MB 00000001 AMI 00000000) Sep 13 01:20:31.026914 kernel: ACPI: SSDT 0x000000008C599EB0 000D8F (v02 INTEL SpsNm 00000002 INTL 20160527) Sep 13 01:20:31.026919 kernel: ACPI: WSMT 0x000000008C59AC40 000028 (v01 SUPERM 01072009 AMI 00010013) Sep 13 01:20:31.026925 kernel: ACPI: EINJ 0x000000008C59AC68 000130 (v01 AMI AMI.EINJ 00000000 AMI. 00000000) Sep 13 01:20:31.026930 kernel: ACPI: ERST 0x000000008C59AD98 000230 (v01 AMIER AMI.ERST 00000000 AMI. 00000000) Sep 13 01:20:31.026936 kernel: ACPI: BERT 0x000000008C59AFC8 000030 (v01 AMI AMI.BERT 00000000 AMI. 00000000) Sep 13 01:20:31.026941 kernel: ACPI: HEST 0x000000008C59AFF8 00027C (v01 AMI AMI.HEST 00000000 AMI. 00000000) Sep 13 01:20:31.026946 kernel: ACPI: SSDT 0x000000008C59B278 000162 (v01 SUPERM SMCCDN 00000000 INTL 20181221) Sep 13 01:20:31.026951 kernel: ACPI: Reserving FACP table memory at [mem 0x8c58b670-0x8c58b783] Sep 13 01:20:31.026956 kernel: ACPI: Reserving DSDT table memory at [mem 0x8c54f268-0x8c58b66b] Sep 13 01:20:31.026961 kernel: ACPI: Reserving FACS table memory at [mem 0x8c66df80-0x8c66dfbf] Sep 13 01:20:31.026966 kernel: ACPI: Reserving APIC table memory at [mem 0x8c58b788-0x8c58b8b3] Sep 13 01:20:31.026971 kernel: ACPI: Reserving FPDT table memory at [mem 0x8c58b8b8-0x8c58b8fb] Sep 13 01:20:31.026977 kernel: ACPI: Reserving FIDT table memory at [mem 0x8c58b900-0x8c58b99b] Sep 13 01:20:31.026983 kernel: ACPI: Reserving MCFG table memory at [mem 0x8c58b9a0-0x8c58b9db] Sep 13 01:20:31.026988 kernel: ACPI: Reserving SPMI table memory at [mem 0x8c58b9e0-0x8c58ba20] Sep 13 01:20:31.026993 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58ba28-0x8c58d543] Sep 13 01:20:31.026998 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58d548-0x8c59070d] Sep 13 01:20:31.027003 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c590710-0x8c592a3a] Sep 13 01:20:31.027008 kernel: ACPI: Reserving HPET table memory at [mem 0x8c592a40-0x8c592a77] Sep 13 01:20:31.027013 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c592a78-0x8c593a25] Sep 13 01:20:31.027018 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c593a28-0x8c59431b] Sep 13 01:20:31.027023 kernel: ACPI: Reserving UEFI table memory at [mem 0x8c594320-0x8c594361] Sep 13 01:20:31.027029 kernel: ACPI: Reserving LPIT table memory at [mem 0x8c594368-0x8c5943fb] Sep 13 01:20:31.027035 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c594400-0x8c596bdd] Sep 13 01:20:31.027040 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c596be0-0x8c5980c1] Sep 13 01:20:31.027045 kernel: ACPI: Reserving DBGP table memory at [mem 0x8c5980c8-0x8c5980fb] Sep 13 01:20:31.027050 kernel: ACPI: Reserving DBG2 table memory at [mem 0x8c598100-0x8c598153] Sep 13 01:20:31.027055 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c598158-0x8c599cbe] Sep 13 01:20:31.027060 kernel: ACPI: Reserving DMAR table memory at [mem 0x8c599cc0-0x8c599d2f] Sep 13 01:20:31.027065 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599d30-0x8c599e73] Sep 13 01:20:31.027070 kernel: ACPI: Reserving TPM2 table memory at [mem 0x8c599e78-0x8c599eab] Sep 13 01:20:31.027076 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599eb0-0x8c59ac3e] Sep 13 01:20:31.027081 kernel: ACPI: Reserving WSMT table memory at [mem 0x8c59ac40-0x8c59ac67] Sep 13 01:20:31.027086 kernel: ACPI: Reserving EINJ table memory at [mem 0x8c59ac68-0x8c59ad97] Sep 13 01:20:31.027092 kernel: ACPI: Reserving ERST table memory at [mem 0x8c59ad98-0x8c59afc7] Sep 13 01:20:31.027097 kernel: ACPI: Reserving BERT table memory at [mem 0x8c59afc8-0x8c59aff7] Sep 13 01:20:31.027102 kernel: ACPI: Reserving HEST table memory at [mem 0x8c59aff8-0x8c59b273] Sep 13 01:20:31.027107 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c59b278-0x8c59b3d9] Sep 13 01:20:31.027112 kernel: No NUMA configuration found Sep 13 01:20:31.027117 kernel: Faking a node at [mem 0x0000000000000000-0x000000086effffff] Sep 13 01:20:31.027122 kernel: NODE_DATA(0) allocated [mem 0x86effa000-0x86effffff] Sep 13 01:20:31.027129 kernel: Zone ranges: Sep 13 01:20:31.027134 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 13 01:20:31.027139 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Sep 13 01:20:31.027144 kernel: Normal [mem 0x0000000100000000-0x000000086effffff] Sep 13 01:20:31.027149 kernel: Movable zone start for each node Sep 13 01:20:31.027154 kernel: Early memory node ranges Sep 13 01:20:31.027159 kernel: node 0: [mem 0x0000000000001000-0x0000000000098fff] Sep 13 01:20:31.027164 kernel: node 0: [mem 0x0000000000100000-0x000000003fffffff] Sep 13 01:20:31.027170 kernel: node 0: [mem 0x0000000040400000-0x00000000825bdfff] Sep 13 01:20:31.027176 kernel: node 0: [mem 0x00000000825c0000-0x000000008afcdfff] Sep 13 01:20:31.027181 kernel: node 0: [mem 0x000000008c0b3000-0x000000008c23bfff] Sep 13 01:20:31.027186 kernel: node 0: [mem 0x000000008eeff000-0x000000008eefffff] Sep 13 01:20:31.027191 kernel: node 0: [mem 0x0000000100000000-0x000000086effffff] Sep 13 01:20:31.027200 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000086effffff] Sep 13 01:20:31.027206 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 13 01:20:31.027211 kernel: On node 0, zone DMA: 103 pages in unavailable ranges Sep 13 01:20:31.027217 kernel: On node 0, zone DMA32: 1024 pages in unavailable ranges Sep 13 01:20:31.027223 kernel: On node 0, zone DMA32: 2 pages in unavailable ranges Sep 13 01:20:31.027228 kernel: On node 0, zone DMA32: 4325 pages in unavailable ranges Sep 13 01:20:31.027234 kernel: On node 0, zone DMA32: 11459 pages in unavailable ranges Sep 13 01:20:31.027239 kernel: On node 0, zone Normal: 4352 pages in unavailable ranges Sep 13 01:20:31.027245 kernel: On node 0, zone Normal: 4096 pages in unavailable ranges Sep 13 01:20:31.027250 kernel: ACPI: PM-Timer IO Port: 0x1808 Sep 13 01:20:31.027256 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Sep 13 01:20:31.027261 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Sep 13 01:20:31.027267 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Sep 13 01:20:31.027273 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Sep 13 01:20:31.027279 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Sep 13 01:20:31.027284 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Sep 13 01:20:31.027290 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Sep 13 01:20:31.027295 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Sep 13 01:20:31.027301 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Sep 13 01:20:31.027306 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Sep 13 01:20:31.027311 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Sep 13 01:20:31.027317 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Sep 13 01:20:31.027323 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Sep 13 01:20:31.027329 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Sep 13 01:20:31.027334 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Sep 13 01:20:31.027342 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Sep 13 01:20:31.027347 kernel: IOAPIC[0]: apic_id 2, version 32, address 0xfec00000, GSI 0-119 Sep 13 01:20:31.027353 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 13 01:20:31.027358 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 13 01:20:31.027364 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 13 01:20:31.027369 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 13 01:20:31.027376 kernel: TSC deadline timer available Sep 13 01:20:31.027382 kernel: smpboot: Allowing 16 CPUs, 0 hotplug CPUs Sep 13 01:20:31.027387 kernel: [mem 0x90000000-0xdfffffff] available for PCI devices Sep 13 01:20:31.027393 kernel: Booting paravirtualized kernel on bare hardware Sep 13 01:20:31.027398 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 13 01:20:31.027404 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Sep 13 01:20:31.027409 kernel: percpu: Embedded 58 pages/cpu s197160 r8192 d32216 u262144 Sep 13 01:20:31.027415 kernel: pcpu-alloc: s197160 r8192 d32216 u262144 alloc=1*2097152 Sep 13 01:20:31.027420 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Sep 13 01:20:31.027427 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:20:31.027433 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 13 01:20:31.027438 kernel: random: crng init done Sep 13 01:20:31.027444 kernel: Dentry cache hash table entries: 4194304 (order: 13, 33554432 bytes, linear) Sep 13 01:20:31.027449 kernel: Inode-cache hash table entries: 2097152 (order: 12, 16777216 bytes, linear) Sep 13 01:20:31.027455 kernel: Fallback order for Node 0: 0 Sep 13 01:20:31.027460 kernel: Built 1 zonelists, mobility grouping on. Total pages: 8232416 Sep 13 01:20:31.027465 kernel: Policy zone: Normal Sep 13 01:20:31.027472 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 13 01:20:31.027477 kernel: software IO TLB: area num 16. Sep 13 01:20:31.027483 kernel: Memory: 32720304K/33452984K available (12288K kernel code, 2293K rwdata, 22744K rodata, 42884K init, 2312K bss, 732420K reserved, 0K cma-reserved) Sep 13 01:20:31.027489 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Sep 13 01:20:31.027494 kernel: ftrace: allocating 37974 entries in 149 pages Sep 13 01:20:31.027500 kernel: ftrace: allocated 149 pages with 4 groups Sep 13 01:20:31.027505 kernel: Dynamic Preempt: voluntary Sep 13 01:20:31.027511 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 13 01:20:31.027517 kernel: rcu: RCU event tracing is enabled. Sep 13 01:20:31.027523 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Sep 13 01:20:31.027529 kernel: Trampoline variant of Tasks RCU enabled. Sep 13 01:20:31.027534 kernel: Rude variant of Tasks RCU enabled. Sep 13 01:20:31.027540 kernel: Tracing variant of Tasks RCU enabled. Sep 13 01:20:31.027545 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 13 01:20:31.027551 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Sep 13 01:20:31.027556 kernel: NR_IRQS: 33024, nr_irqs: 2184, preallocated irqs: 16 Sep 13 01:20:31.027562 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 13 01:20:31.027567 kernel: Console: colour dummy device 80x25 Sep 13 01:20:31.027573 kernel: printk: console [tty0] enabled Sep 13 01:20:31.027579 kernel: printk: console [ttyS1] enabled Sep 13 01:20:31.027585 kernel: ACPI: Core revision 20230628 Sep 13 01:20:31.027590 kernel: hpet: HPET dysfunctional in PC10. Force disabled. Sep 13 01:20:31.027596 kernel: APIC: Switch to symmetric I/O mode setup Sep 13 01:20:31.027601 kernel: DMAR: Host address width 39 Sep 13 01:20:31.027606 kernel: DMAR: DRHD base: 0x000000fed91000 flags: 0x1 Sep 13 01:20:31.027612 kernel: DMAR: dmar0: reg_base_addr fed91000 ver 1:0 cap d2008c40660462 ecap f050da Sep 13 01:20:31.027617 kernel: DMAR: RMRR base: 0x0000008cf19000 end: 0x0000008d162fff Sep 13 01:20:31.027623 kernel: DMAR-IR: IOAPIC id 2 under DRHD base 0xfed91000 IOMMU 0 Sep 13 01:20:31.027629 kernel: DMAR-IR: HPET id 0 under DRHD base 0xfed91000 Sep 13 01:20:31.027635 kernel: DMAR-IR: Queued invalidation will be enabled to support x2apic and Intr-remapping. Sep 13 01:20:31.027640 kernel: DMAR-IR: Enabled IRQ remapping in x2apic mode Sep 13 01:20:31.027646 kernel: x2apic enabled Sep 13 01:20:31.027651 kernel: APIC: Switched APIC routing to: cluster x2apic Sep 13 01:20:31.027657 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x3101f59f5e6, max_idle_ns: 440795259996 ns Sep 13 01:20:31.027663 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 6799.81 BogoMIPS (lpj=3399906) Sep 13 01:20:31.027668 kernel: CPU0: Thermal monitoring enabled (TM1) Sep 13 01:20:31.027673 kernel: process: using mwait in idle threads Sep 13 01:20:31.027680 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Sep 13 01:20:31.027685 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Sep 13 01:20:31.027691 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 13 01:20:31.027696 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Sep 13 01:20:31.027701 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Sep 13 01:20:31.027707 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Sep 13 01:20:31.027712 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Sep 13 01:20:31.027718 kernel: RETBleed: Mitigation: Enhanced IBRS Sep 13 01:20:31.027723 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 13 01:20:31.027729 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 13 01:20:31.027734 kernel: TAA: Mitigation: TSX disabled Sep 13 01:20:31.027740 kernel: MMIO Stale Data: Mitigation: Clear CPU buffers Sep 13 01:20:31.027746 kernel: SRBDS: Mitigation: Microcode Sep 13 01:20:31.027751 kernel: GDS: Mitigation: Microcode Sep 13 01:20:31.027757 kernel: active return thunk: its_return_thunk Sep 13 01:20:31.027762 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 13 01:20:31.027767 kernel: VMSCAPE: Mitigation: IBPB before exit to userspace Sep 13 01:20:31.027773 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 13 01:20:31.027778 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 13 01:20:31.027784 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 13 01:20:31.027789 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Sep 13 01:20:31.027794 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Sep 13 01:20:31.027801 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 13 01:20:31.027806 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Sep 13 01:20:31.027812 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Sep 13 01:20:31.027817 kernel: x86/fpu: Enabled xstate features 0x1f, context size is 960 bytes, using 'compacted' format. Sep 13 01:20:31.027823 kernel: Freeing SMP alternatives memory: 32K Sep 13 01:20:31.027828 kernel: pid_max: default: 32768 minimum: 301 Sep 13 01:20:31.027834 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 13 01:20:31.027839 kernel: landlock: Up and running. Sep 13 01:20:31.027845 kernel: SELinux: Initializing. Sep 13 01:20:31.027850 kernel: Mount-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.027856 kernel: Mountpoint-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.027862 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Sep 13 01:20:31.027868 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:20:31.027873 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:20:31.027879 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:20:31.027884 kernel: Performance Events: PEBS fmt3+, Skylake events, 32-deep LBR, full-width counters, Intel PMU driver. Sep 13 01:20:31.027890 kernel: ... version: 4 Sep 13 01:20:31.027896 kernel: ... bit width: 48 Sep 13 01:20:31.027901 kernel: ... generic registers: 4 Sep 13 01:20:31.027906 kernel: ... value mask: 0000ffffffffffff Sep 13 01:20:31.027913 kernel: ... max period: 00007fffffffffff Sep 13 01:20:31.027919 kernel: ... fixed-purpose events: 3 Sep 13 01:20:31.027924 kernel: ... event mask: 000000070000000f Sep 13 01:20:31.027930 kernel: signal: max sigframe size: 2032 Sep 13 01:20:31.027935 kernel: Estimated ratio of average max frequency by base frequency (times 1024): 1445 Sep 13 01:20:31.027941 kernel: rcu: Hierarchical SRCU implementation. Sep 13 01:20:31.027946 kernel: rcu: Max phase no-delay instances is 400. Sep 13 01:20:31.027952 kernel: NMI watchdog: Enabled. Permanently consumes one hw-PMU counter. Sep 13 01:20:31.027957 kernel: smp: Bringing up secondary CPUs ... Sep 13 01:20:31.027963 kernel: smpboot: x86: Booting SMP configuration: Sep 13 01:20:31.027969 kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6 #7 #8 #9 #10 #11 #12 #13 #14 #15 Sep 13 01:20:31.027975 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 13 01:20:31.027981 kernel: smp: Brought up 1 node, 16 CPUs Sep 13 01:20:31.027986 kernel: smpboot: Max logical packages: 1 Sep 13 01:20:31.027992 kernel: smpboot: Total of 16 processors activated (108796.99 BogoMIPS) Sep 13 01:20:31.027997 kernel: devtmpfs: initialized Sep 13 01:20:31.028003 kernel: x86/mm: Memory block size: 128MB Sep 13 01:20:31.028008 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x825be000-0x825befff] (4096 bytes) Sep 13 01:20:31.028015 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x8c23c000-0x8c66dfff] (4399104 bytes) Sep 13 01:20:31.028020 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 13 01:20:31.028026 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Sep 13 01:20:31.028031 kernel: pinctrl core: initialized pinctrl subsystem Sep 13 01:20:31.028037 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 13 01:20:31.028042 kernel: audit: initializing netlink subsys (disabled) Sep 13 01:20:31.028048 kernel: audit: type=2000 audit(1757726425.039:1): state=initialized audit_enabled=0 res=1 Sep 13 01:20:31.028053 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 13 01:20:31.028059 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 13 01:20:31.028065 kernel: cpuidle: using governor menu Sep 13 01:20:31.028070 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 13 01:20:31.028076 kernel: dca service started, version 1.12.1 Sep 13 01:20:31.028081 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Sep 13 01:20:31.028087 kernel: PCI: Using configuration type 1 for base access Sep 13 01:20:31.028092 kernel: ENERGY_PERF_BIAS: Set to 'normal', was 'performance' Sep 13 01:20:31.028098 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 13 01:20:31.028103 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 13 01:20:31.028109 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 13 01:20:31.028115 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 13 01:20:31.028121 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 13 01:20:31.028126 kernel: ACPI: Added _OSI(Module Device) Sep 13 01:20:31.028132 kernel: ACPI: Added _OSI(Processor Device) Sep 13 01:20:31.028137 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 13 01:20:31.028143 kernel: ACPI: 12 ACPI AML tables successfully acquired and loaded Sep 13 01:20:31.028148 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028154 kernel: ACPI: SSDT 0xFFFF98DD41AF2400 000400 (v02 PmRef Cpu0Cst 00003001 INTL 20160527) Sep 13 01:20:31.028159 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028166 kernel: ACPI: SSDT 0xFFFF98DD41AEC800 000683 (v02 PmRef Cpu0Ist 00003000 INTL 20160527) Sep 13 01:20:31.028171 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028177 kernel: ACPI: SSDT 0xFFFF98DD40247100 0000F4 (v02 PmRef Cpu0Psd 00003000 INTL 20160527) Sep 13 01:20:31.028182 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028187 kernel: ACPI: SSDT 0xFFFF98DD41E5F800 0005FC (v02 PmRef ApIst 00003000 INTL 20160527) Sep 13 01:20:31.028193 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028198 kernel: ACPI: SSDT 0xFFFF98DD4012E000 000AB0 (v02 PmRef ApPsd 00003000 INTL 20160527) Sep 13 01:20:31.028204 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028209 kernel: ACPI: SSDT 0xFFFF98DD41AF4C00 00030A (v02 PmRef ApCst 00003000 INTL 20160527) Sep 13 01:20:31.028215 kernel: ACPI: _OSC evaluated successfully for all CPUs Sep 13 01:20:31.028221 kernel: ACPI: Interpreter enabled Sep 13 01:20:31.028226 kernel: ACPI: PM: (supports S0 S5) Sep 13 01:20:31.028232 kernel: ACPI: Using IOAPIC for interrupt routing Sep 13 01:20:31.028237 kernel: HEST: Enabling Firmware First mode for corrected errors. Sep 13 01:20:31.028243 kernel: mce: [Firmware Bug]: Ignoring request to disable invalid MCA bank 14. Sep 13 01:20:31.028248 kernel: HEST: Table parsing has been initialized. Sep 13 01:20:31.028254 kernel: GHES: APEI firmware first mode is enabled by APEI bit and WHEA _OSC. Sep 13 01:20:31.028259 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 13 01:20:31.028265 kernel: PCI: Ignoring E820 reservations for host bridge windows Sep 13 01:20:31.028271 kernel: ACPI: Enabled 9 GPEs in block 00 to 7F Sep 13 01:20:31.028277 kernel: ACPI: \_SB_.PCI0.XDCI.USBC: New power resource Sep 13 01:20:31.028282 kernel: ACPI: \_SB_.PCI0.SAT0.VOL0.V0PR: New power resource Sep 13 01:20:31.028288 kernel: ACPI: \_SB_.PCI0.SAT0.VOL1.V1PR: New power resource Sep 13 01:20:31.028293 kernel: ACPI: \_SB_.PCI0.SAT0.VOL2.V2PR: New power resource Sep 13 01:20:31.028299 kernel: ACPI: \_SB_.PCI0.CNVW.WRST: New power resource Sep 13 01:20:31.028304 kernel: ACPI: \_TZ_.FN00: New power resource Sep 13 01:20:31.028310 kernel: ACPI: \_TZ_.FN01: New power resource Sep 13 01:20:31.028315 kernel: ACPI: \_TZ_.FN02: New power resource Sep 13 01:20:31.028322 kernel: ACPI: \_TZ_.FN03: New power resource Sep 13 01:20:31.028327 kernel: ACPI: \_TZ_.FN04: New power resource Sep 13 01:20:31.028333 kernel: ACPI: \PIN_: New power resource Sep 13 01:20:31.028340 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-fe]) Sep 13 01:20:31.028415 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 13 01:20:31.028470 kernel: acpi PNP0A08:00: _OSC: platform does not support [AER] Sep 13 01:20:31.028520 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability LTR] Sep 13 01:20:31.028530 kernel: PCI host bridge to bus 0000:00 Sep 13 01:20:31.028580 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 13 01:20:31.028624 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 13 01:20:31.028668 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 13 01:20:31.028711 kernel: pci_bus 0000:00: root bus resource [mem 0x90000000-0xdfffffff window] Sep 13 01:20:31.028753 kernel: pci_bus 0000:00: root bus resource [mem 0xfc800000-0xfe7fffff window] Sep 13 01:20:31.028796 kernel: pci_bus 0000:00: root bus resource [bus 00-fe] Sep 13 01:20:31.028859 kernel: pci 0000:00:00.0: [8086:3e31] type 00 class 0x060000 Sep 13 01:20:31.028916 kernel: pci 0000:00:01.0: [8086:1901] type 01 class 0x060400 Sep 13 01:20:31.028966 kernel: pci 0000:00:01.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.029019 kernel: pci 0000:00:08.0: [8086:1911] type 00 class 0x088000 Sep 13 01:20:31.029069 kernel: pci 0000:00:08.0: reg 0x10: [mem 0x9551f000-0x9551ffff 64bit] Sep 13 01:20:31.029120 kernel: pci 0000:00:12.0: [8086:a379] type 00 class 0x118000 Sep 13 01:20:31.029172 kernel: pci 0000:00:12.0: reg 0x10: [mem 0x9551e000-0x9551efff 64bit] Sep 13 01:20:31.029225 kernel: pci 0000:00:14.0: [8086:a36d] type 00 class 0x0c0330 Sep 13 01:20:31.029274 kernel: pci 0000:00:14.0: reg 0x10: [mem 0x95500000-0x9550ffff 64bit] Sep 13 01:20:31.029322 kernel: pci 0000:00:14.0: PME# supported from D3hot D3cold Sep 13 01:20:31.029380 kernel: pci 0000:00:14.2: [8086:a36f] type 00 class 0x050000 Sep 13 01:20:31.029429 kernel: pci 0000:00:14.2: reg 0x10: [mem 0x95512000-0x95513fff 64bit] Sep 13 01:20:31.029481 kernel: pci 0000:00:14.2: reg 0x18: [mem 0x9551d000-0x9551dfff 64bit] Sep 13 01:20:31.029533 kernel: pci 0000:00:15.0: [8086:a368] type 00 class 0x0c8000 Sep 13 01:20:31.029582 kernel: pci 0000:00:15.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Sep 13 01:20:31.029638 kernel: pci 0000:00:15.1: [8086:a369] type 00 class 0x0c8000 Sep 13 01:20:31.029688 kernel: pci 0000:00:15.1: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Sep 13 01:20:31.029741 kernel: pci 0000:00:16.0: [8086:a360] type 00 class 0x078000 Sep 13 01:20:31.029792 kernel: pci 0000:00:16.0: reg 0x10: [mem 0x9551a000-0x9551afff 64bit] Sep 13 01:20:31.029843 kernel: pci 0000:00:16.0: PME# supported from D3hot Sep 13 01:20:31.029903 kernel: pci 0000:00:16.1: [8086:a361] type 00 class 0x078000 Sep 13 01:20:31.029955 kernel: pci 0000:00:16.1: reg 0x10: [mem 0x95519000-0x95519fff 64bit] Sep 13 01:20:31.030003 kernel: pci 0000:00:16.1: PME# supported from D3hot Sep 13 01:20:31.030055 kernel: pci 0000:00:16.4: [8086:a364] type 00 class 0x078000 Sep 13 01:20:31.030104 kernel: pci 0000:00:16.4: reg 0x10: [mem 0x95518000-0x95518fff 64bit] Sep 13 01:20:31.030155 kernel: pci 0000:00:16.4: PME# supported from D3hot Sep 13 01:20:31.030208 kernel: pci 0000:00:17.0: [8086:a352] type 00 class 0x010601 Sep 13 01:20:31.030258 kernel: pci 0000:00:17.0: reg 0x10: [mem 0x95510000-0x95511fff] Sep 13 01:20:31.030306 kernel: pci 0000:00:17.0: reg 0x14: [mem 0x95517000-0x955170ff] Sep 13 01:20:31.030357 kernel: pci 0000:00:17.0: reg 0x18: [io 0x6050-0x6057] Sep 13 01:20:31.030405 kernel: pci 0000:00:17.0: reg 0x1c: [io 0x6040-0x6043] Sep 13 01:20:31.030454 kernel: pci 0000:00:17.0: reg 0x20: [io 0x6020-0x603f] Sep 13 01:20:31.030506 kernel: pci 0000:00:17.0: reg 0x24: [mem 0x95516000-0x955167ff] Sep 13 01:20:31.030554 kernel: pci 0000:00:17.0: PME# supported from D3hot Sep 13 01:20:31.030610 kernel: pci 0000:00:1b.0: [8086:a340] type 01 class 0x060400 Sep 13 01:20:31.030659 kernel: pci 0000:00:1b.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.030718 kernel: pci 0000:00:1b.4: [8086:a32c] type 01 class 0x060400 Sep 13 01:20:31.030768 kernel: pci 0000:00:1b.4: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.030821 kernel: pci 0000:00:1b.5: [8086:a32d] type 01 class 0x060400 Sep 13 01:20:31.030870 kernel: pci 0000:00:1b.5: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.030924 kernel: pci 0000:00:1c.0: [8086:a338] type 01 class 0x060400 Sep 13 01:20:31.030975 kernel: pci 0000:00:1c.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.031033 kernel: pci 0000:00:1c.3: [8086:a33b] type 01 class 0x060400 Sep 13 01:20:31.031082 kernel: pci 0000:00:1c.3: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.031136 kernel: pci 0000:00:1e.0: [8086:a328] type 00 class 0x078000 Sep 13 01:20:31.031184 kernel: pci 0000:00:1e.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Sep 13 01:20:31.031238 kernel: pci 0000:00:1f.0: [8086:a309] type 00 class 0x060100 Sep 13 01:20:31.031290 kernel: pci 0000:00:1f.4: [8086:a323] type 00 class 0x0c0500 Sep 13 01:20:31.031346 kernel: pci 0000:00:1f.4: reg 0x10: [mem 0x95514000-0x955140ff 64bit] Sep 13 01:20:31.031396 kernel: pci 0000:00:1f.4: reg 0x20: [io 0xefa0-0xefbf] Sep 13 01:20:31.031451 kernel: pci 0000:00:1f.5: [8086:a324] type 00 class 0x0c8000 Sep 13 01:20:31.031500 kernel: pci 0000:00:1f.5: reg 0x10: [mem 0xfe010000-0xfe010fff] Sep 13 01:20:31.031555 kernel: pci 0000:01:00.0: [15b3:1015] type 00 class 0x020000 Sep 13 01:20:31.031606 kernel: pci 0000:01:00.0: reg 0x10: [mem 0x92000000-0x93ffffff 64bit pref] Sep 13 01:20:31.031659 kernel: pci 0000:01:00.0: reg 0x30: [mem 0x95200000-0x952fffff pref] Sep 13 01:20:31.031709 kernel: pci 0000:01:00.0: PME# supported from D3cold Sep 13 01:20:31.031759 kernel: pci 0000:01:00.0: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Sep 13 01:20:31.031810 kernel: pci 0000:01:00.0: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Sep 13 01:20:31.031865 kernel: pci 0000:01:00.1: [15b3:1015] type 00 class 0x020000 Sep 13 01:20:31.031915 kernel: pci 0000:01:00.1: reg 0x10: [mem 0x90000000-0x91ffffff 64bit pref] Sep 13 01:20:31.031966 kernel: pci 0000:01:00.1: reg 0x30: [mem 0x95100000-0x951fffff pref] Sep 13 01:20:31.032017 kernel: pci 0000:01:00.1: PME# supported from D3cold Sep 13 01:20:31.032068 kernel: pci 0000:01:00.1: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Sep 13 01:20:31.032117 kernel: pci 0000:01:00.1: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Sep 13 01:20:31.032168 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Sep 13 01:20:31.032217 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Sep 13 01:20:31.032267 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Sep 13 01:20:31.032315 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Sep 13 01:20:31.032375 kernel: pci 0000:03:00.0: working around ROM BAR overlap defect Sep 13 01:20:31.032429 kernel: pci 0000:03:00.0: [8086:1533] type 00 class 0x020000 Sep 13 01:20:31.032480 kernel: pci 0000:03:00.0: reg 0x10: [mem 0x95400000-0x9547ffff] Sep 13 01:20:31.032530 kernel: pci 0000:03:00.0: reg 0x18: [io 0x5000-0x501f] Sep 13 01:20:31.032579 kernel: pci 0000:03:00.0: reg 0x1c: [mem 0x95480000-0x95483fff] Sep 13 01:20:31.032631 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.032681 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Sep 13 01:20:31.032731 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Sep 13 01:20:31.032783 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Sep 13 01:20:31.032839 kernel: pci 0000:04:00.0: working around ROM BAR overlap defect Sep 13 01:20:31.032891 kernel: pci 0000:04:00.0: [8086:1533] type 00 class 0x020000 Sep 13 01:20:31.032941 kernel: pci 0000:04:00.0: reg 0x10: [mem 0x95300000-0x9537ffff] Sep 13 01:20:31.032993 kernel: pci 0000:04:00.0: reg 0x18: [io 0x4000-0x401f] Sep 13 01:20:31.033043 kernel: pci 0000:04:00.0: reg 0x1c: [mem 0x95380000-0x95383fff] Sep 13 01:20:31.033094 kernel: pci 0000:04:00.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.033146 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Sep 13 01:20:31.033196 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Sep 13 01:20:31.033245 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Sep 13 01:20:31.033295 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Sep 13 01:20:31.033355 kernel: pci 0000:06:00.0: [1a03:1150] type 01 class 0x060400 Sep 13 01:20:31.033406 kernel: pci 0000:06:00.0: enabling Extended Tags Sep 13 01:20:31.033457 kernel: pci 0000:06:00.0: supports D1 D2 Sep 13 01:20:31.033506 kernel: pci 0000:06:00.0: PME# supported from D0 D1 D2 D3hot D3cold Sep 13 01:20:31.033559 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Sep 13 01:20:31.033608 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.033658 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.033713 kernel: pci_bus 0000:07: extended config space not accessible Sep 13 01:20:31.033771 kernel: pci 0000:07:00.0: [1a03:2000] type 00 class 0x030000 Sep 13 01:20:31.033824 kernel: pci 0000:07:00.0: reg 0x10: [mem 0x94000000-0x94ffffff] Sep 13 01:20:31.033875 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x95000000-0x9501ffff] Sep 13 01:20:31.033931 kernel: pci 0000:07:00.0: reg 0x18: [io 0x3000-0x307f] Sep 13 01:20:31.033982 kernel: pci 0000:07:00.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 13 01:20:31.034035 kernel: pci 0000:07:00.0: supports D1 D2 Sep 13 01:20:31.034088 kernel: pci 0000:07:00.0: PME# supported from D0 D1 D2 D3hot D3cold Sep 13 01:20:31.034138 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Sep 13 01:20:31.034188 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.034238 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.034246 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 0 Sep 13 01:20:31.034254 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 1 Sep 13 01:20:31.034260 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 0 Sep 13 01:20:31.034267 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 0 Sep 13 01:20:31.034273 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 0 Sep 13 01:20:31.034279 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 0 Sep 13 01:20:31.034285 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 0 Sep 13 01:20:31.034290 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 0 Sep 13 01:20:31.034296 kernel: iommu: Default domain type: Translated Sep 13 01:20:31.034302 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 13 01:20:31.034309 kernel: PCI: Using ACPI for IRQ routing Sep 13 01:20:31.034315 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 13 01:20:31.034320 kernel: e820: reserve RAM buffer [mem 0x00099800-0x0009ffff] Sep 13 01:20:31.034326 kernel: e820: reserve RAM buffer [mem 0x825be000-0x83ffffff] Sep 13 01:20:31.034332 kernel: e820: reserve RAM buffer [mem 0x8afce000-0x8bffffff] Sep 13 01:20:31.034339 kernel: e820: reserve RAM buffer [mem 0x8c23c000-0x8fffffff] Sep 13 01:20:31.034345 kernel: e820: reserve RAM buffer [mem 0x8ef00000-0x8fffffff] Sep 13 01:20:31.034351 kernel: e820: reserve RAM buffer [mem 0x86f000000-0x86fffffff] Sep 13 01:20:31.034403 kernel: pci 0000:07:00.0: vgaarb: setting as boot VGA device Sep 13 01:20:31.034457 kernel: pci 0000:07:00.0: vgaarb: bridge control possible Sep 13 01:20:31.034510 kernel: pci 0000:07:00.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 13 01:20:31.034518 kernel: vgaarb: loaded Sep 13 01:20:31.034524 kernel: clocksource: Switched to clocksource tsc-early Sep 13 01:20:31.034530 kernel: VFS: Disk quotas dquot_6.6.0 Sep 13 01:20:31.034536 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 13 01:20:31.034542 kernel: pnp: PnP ACPI init Sep 13 01:20:31.034592 kernel: system 00:00: [mem 0x40000000-0x403fffff] has been reserved Sep 13 01:20:31.034643 kernel: pnp 00:02: [dma 0 disabled] Sep 13 01:20:31.034693 kernel: pnp 00:03: [dma 0 disabled] Sep 13 01:20:31.034745 kernel: system 00:04: [io 0x0680-0x069f] has been reserved Sep 13 01:20:31.034791 kernel: system 00:04: [io 0x164e-0x164f] has been reserved Sep 13 01:20:31.034839 kernel: system 00:05: [mem 0xfed10000-0xfed17fff] has been reserved Sep 13 01:20:31.034884 kernel: system 00:05: [mem 0xfed18000-0xfed18fff] has been reserved Sep 13 01:20:31.034931 kernel: system 00:05: [mem 0xfed19000-0xfed19fff] has been reserved Sep 13 01:20:31.034976 kernel: system 00:05: [mem 0xe0000000-0xefffffff] has been reserved Sep 13 01:20:31.035020 kernel: system 00:05: [mem 0xfed20000-0xfed3ffff] has been reserved Sep 13 01:20:31.035068 kernel: system 00:05: [mem 0xfed90000-0xfed93fff] could not be reserved Sep 13 01:20:31.035113 kernel: system 00:05: [mem 0xfed45000-0xfed8ffff] has been reserved Sep 13 01:20:31.035158 kernel: system 00:05: [mem 0xfee00000-0xfeefffff] could not be reserved Sep 13 01:20:31.035206 kernel: system 00:06: [io 0x1800-0x18fe] could not be reserved Sep 13 01:20:31.035255 kernel: system 00:06: [mem 0xfd000000-0xfd69ffff] has been reserved Sep 13 01:20:31.035299 kernel: system 00:06: [mem 0xfd6c0000-0xfd6cffff] has been reserved Sep 13 01:20:31.035346 kernel: system 00:06: [mem 0xfd6f0000-0xfdffffff] has been reserved Sep 13 01:20:31.035391 kernel: system 00:06: [mem 0xfe000000-0xfe01ffff] could not be reserved Sep 13 01:20:31.035435 kernel: system 00:06: [mem 0xfe200000-0xfe7fffff] has been reserved Sep 13 01:20:31.035480 kernel: system 00:06: [mem 0xff000000-0xffffffff] has been reserved Sep 13 01:20:31.035528 kernel: system 00:07: [io 0x2000-0x20fe] has been reserved Sep 13 01:20:31.035538 kernel: pnp: PnP ACPI: found 9 devices Sep 13 01:20:31.035544 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 13 01:20:31.035550 kernel: NET: Registered PF_INET protocol family Sep 13 01:20:31.035556 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 01:20:31.035562 kernel: tcp_listen_portaddr_hash hash table entries: 16384 (order: 6, 262144 bytes, linear) Sep 13 01:20:31.035568 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 13 01:20:31.035574 kernel: TCP established hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 01:20:31.035579 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Sep 13 01:20:31.035586 kernel: TCP: Hash tables configured (established 262144 bind 65536) Sep 13 01:20:31.035592 kernel: UDP hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.035598 kernel: UDP-Lite hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.035604 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 13 01:20:31.035610 kernel: NET: Registered PF_XDP protocol family Sep 13 01:20:31.035659 kernel: pci 0000:00:15.0: BAR 0: assigned [mem 0x95515000-0x95515fff 64bit] Sep 13 01:20:31.035709 kernel: pci 0000:00:15.1: BAR 0: assigned [mem 0x9551b000-0x9551bfff 64bit] Sep 13 01:20:31.035760 kernel: pci 0000:00:1e.0: BAR 0: assigned [mem 0x9551c000-0x9551cfff 64bit] Sep 13 01:20:31.035811 kernel: pci 0000:01:00.0: BAR 7: no space for [mem size 0x00800000 64bit pref] Sep 13 01:20:31.035866 kernel: pci 0000:01:00.0: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Sep 13 01:20:31.035917 kernel: pci 0000:01:00.1: BAR 7: no space for [mem size 0x00800000 64bit pref] Sep 13 01:20:31.035968 kernel: pci 0000:01:00.1: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Sep 13 01:20:31.036018 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Sep 13 01:20:31.036067 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Sep 13 01:20:31.036116 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Sep 13 01:20:31.036168 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Sep 13 01:20:31.036220 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Sep 13 01:20:31.036268 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Sep 13 01:20:31.036318 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Sep 13 01:20:31.036370 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Sep 13 01:20:31.036420 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Sep 13 01:20:31.036470 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Sep 13 01:20:31.036520 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Sep 13 01:20:31.036569 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Sep 13 01:20:31.036619 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.036670 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.036719 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Sep 13 01:20:31.036769 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.036819 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.036865 kernel: pci_bus 0000:00: Some PCI device resources are unassigned, try booting with pci=realloc Sep 13 01:20:31.036911 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 13 01:20:31.036956 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 13 01:20:31.036998 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 13 01:20:31.037042 kernel: pci_bus 0000:00: resource 7 [mem 0x90000000-0xdfffffff window] Sep 13 01:20:31.037084 kernel: pci_bus 0000:00: resource 8 [mem 0xfc800000-0xfe7fffff window] Sep 13 01:20:31.037134 kernel: pci_bus 0000:01: resource 1 [mem 0x95100000-0x952fffff] Sep 13 01:20:31.037180 kernel: pci_bus 0000:01: resource 2 [mem 0x90000000-0x93ffffff 64bit pref] Sep 13 01:20:31.037232 kernel: pci_bus 0000:03: resource 0 [io 0x5000-0x5fff] Sep 13 01:20:31.037277 kernel: pci_bus 0000:03: resource 1 [mem 0x95400000-0x954fffff] Sep 13 01:20:31.037330 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Sep 13 01:20:31.037377 kernel: pci_bus 0000:04: resource 1 [mem 0x95300000-0x953fffff] Sep 13 01:20:31.037427 kernel: pci_bus 0000:06: resource 0 [io 0x3000-0x3fff] Sep 13 01:20:31.037473 kernel: pci_bus 0000:06: resource 1 [mem 0x94000000-0x950fffff] Sep 13 01:20:31.037520 kernel: pci_bus 0000:07: resource 0 [io 0x3000-0x3fff] Sep 13 01:20:31.037569 kernel: pci_bus 0000:07: resource 1 [mem 0x94000000-0x950fffff] Sep 13 01:20:31.037577 kernel: PCI: CLS 64 bytes, default 64 Sep 13 01:20:31.037583 kernel: DMAR: No ATSR found Sep 13 01:20:31.037589 kernel: DMAR: No SATC found Sep 13 01:20:31.037595 kernel: DMAR: dmar0: Using Queued invalidation Sep 13 01:20:31.037644 kernel: pci 0000:00:00.0: Adding to iommu group 0 Sep 13 01:20:31.037694 kernel: pci 0000:00:01.0: Adding to iommu group 1 Sep 13 01:20:31.037744 kernel: pci 0000:00:08.0: Adding to iommu group 2 Sep 13 01:20:31.037796 kernel: pci 0000:00:12.0: Adding to iommu group 3 Sep 13 01:20:31.037845 kernel: pci 0000:00:14.0: Adding to iommu group 4 Sep 13 01:20:31.037896 kernel: pci 0000:00:14.2: Adding to iommu group 4 Sep 13 01:20:31.037944 kernel: pci 0000:00:15.0: Adding to iommu group 5 Sep 13 01:20:31.037993 kernel: pci 0000:00:15.1: Adding to iommu group 5 Sep 13 01:20:31.038042 kernel: pci 0000:00:16.0: Adding to iommu group 6 Sep 13 01:20:31.038091 kernel: pci 0000:00:16.1: Adding to iommu group 6 Sep 13 01:20:31.038140 kernel: pci 0000:00:16.4: Adding to iommu group 6 Sep 13 01:20:31.038191 kernel: pci 0000:00:17.0: Adding to iommu group 7 Sep 13 01:20:31.038239 kernel: pci 0000:00:1b.0: Adding to iommu group 8 Sep 13 01:20:31.038288 kernel: pci 0000:00:1b.4: Adding to iommu group 9 Sep 13 01:20:31.038336 kernel: pci 0000:00:1b.5: Adding to iommu group 10 Sep 13 01:20:31.038390 kernel: pci 0000:00:1c.0: Adding to iommu group 11 Sep 13 01:20:31.038440 kernel: pci 0000:00:1c.3: Adding to iommu group 12 Sep 13 01:20:31.038488 kernel: pci 0000:00:1e.0: Adding to iommu group 13 Sep 13 01:20:31.038538 kernel: pci 0000:00:1f.0: Adding to iommu group 14 Sep 13 01:20:31.038589 kernel: pci 0000:00:1f.4: Adding to iommu group 14 Sep 13 01:20:31.038639 kernel: pci 0000:00:1f.5: Adding to iommu group 14 Sep 13 01:20:31.038689 kernel: pci 0000:01:00.0: Adding to iommu group 1 Sep 13 01:20:31.038740 kernel: pci 0000:01:00.1: Adding to iommu group 1 Sep 13 01:20:31.038791 kernel: pci 0000:03:00.0: Adding to iommu group 15 Sep 13 01:20:31.038841 kernel: pci 0000:04:00.0: Adding to iommu group 16 Sep 13 01:20:31.038892 kernel: pci 0000:06:00.0: Adding to iommu group 17 Sep 13 01:20:31.038946 kernel: pci 0000:07:00.0: Adding to iommu group 17 Sep 13 01:20:31.038956 kernel: DMAR: Intel(R) Virtualization Technology for Directed I/O Sep 13 01:20:31.038962 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Sep 13 01:20:31.038968 kernel: software IO TLB: mapped [mem 0x0000000086fce000-0x000000008afce000] (64MB) Sep 13 01:20:31.038974 kernel: RAPL PMU: API unit is 2^-32 Joules, 3 fixed counters, 655360 ms ovfl timer Sep 13 01:20:31.038980 kernel: RAPL PMU: hw unit of domain pp0-core 2^-14 Joules Sep 13 01:20:31.038986 kernel: RAPL PMU: hw unit of domain package 2^-14 Joules Sep 13 01:20:31.038992 kernel: RAPL PMU: hw unit of domain dram 2^-14 Joules Sep 13 01:20:31.039044 kernel: platform rtc_cmos: registered platform RTC device (no PNP device found) Sep 13 01:20:31.039053 kernel: Initialise system trusted keyrings Sep 13 01:20:31.039061 kernel: workingset: timestamp_bits=39 max_order=23 bucket_order=0 Sep 13 01:20:31.039066 kernel: Key type asymmetric registered Sep 13 01:20:31.039072 kernel: Asymmetric key parser 'x509' registered Sep 13 01:20:31.039078 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 13 01:20:31.039084 kernel: io scheduler mq-deadline registered Sep 13 01:20:31.039089 kernel: io scheduler kyber registered Sep 13 01:20:31.039095 kernel: io scheduler bfq registered Sep 13 01:20:31.039143 kernel: pcieport 0000:00:01.0: PME: Signaling with IRQ 121 Sep 13 01:20:31.039193 kernel: pcieport 0000:00:1b.0: PME: Signaling with IRQ 122 Sep 13 01:20:31.039245 kernel: pcieport 0000:00:1b.4: PME: Signaling with IRQ 123 Sep 13 01:20:31.039294 kernel: pcieport 0000:00:1b.5: PME: Signaling with IRQ 124 Sep 13 01:20:31.039346 kernel: pcieport 0000:00:1c.0: PME: Signaling with IRQ 125 Sep 13 01:20:31.039397 kernel: pcieport 0000:00:1c.3: PME: Signaling with IRQ 126 Sep 13 01:20:31.039450 kernel: thermal LNXTHERM:00: registered as thermal_zone0 Sep 13 01:20:31.039459 kernel: ACPI: thermal: Thermal Zone [TZ00] (28 C) Sep 13 01:20:31.039465 kernel: ERST: Error Record Serialization Table (ERST) support is initialized. Sep 13 01:20:31.039473 kernel: pstore: Using crash dump compression: deflate Sep 13 01:20:31.039479 kernel: pstore: Registered erst as persistent store backend Sep 13 01:20:31.039485 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 13 01:20:31.039491 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 13 01:20:31.039497 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 13 01:20:31.039503 kernel: 00:03: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Sep 13 01:20:31.039509 kernel: hpet_acpi_add: no address or irqs in _CRS Sep 13 01:20:31.039560 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x1B, rev-id 16) Sep 13 01:20:31.039570 kernel: i8042: PNP: No PS/2 controller found. Sep 13 01:20:31.039615 kernel: rtc_cmos rtc_cmos: RTC can wake from S4 Sep 13 01:20:31.039661 kernel: rtc_cmos rtc_cmos: registered as rtc0 Sep 13 01:20:31.039706 kernel: rtc_cmos rtc_cmos: setting system clock to 2025-09-13T01:20:29 UTC (1757726429) Sep 13 01:20:31.039752 kernel: rtc_cmos rtc_cmos: alarms up to one month, y3k, 114 bytes nvram Sep 13 01:20:31.039760 kernel: intel_pstate: Intel P-state driver initializing Sep 13 01:20:31.039767 kernel: intel_pstate: Disabling energy efficiency optimization Sep 13 01:20:31.039772 kernel: intel_pstate: HWP enabled Sep 13 01:20:31.039780 kernel: vesafb: mode is 1024x768x8, linelength=1024, pages=0 Sep 13 01:20:31.039786 kernel: vesafb: scrolling: redraw Sep 13 01:20:31.039792 kernel: vesafb: Pseudocolor: size=0:8:8:8, shift=0:0:0:0 Sep 13 01:20:31.039798 kernel: vesafb: framebuffer at 0x94000000, mapped to 0x0000000067ea91ba, using 768k, total 768k Sep 13 01:20:31.039803 kernel: Console: switching to colour frame buffer device 128x48 Sep 13 01:20:31.039809 kernel: fb0: VESA VGA frame buffer device Sep 13 01:20:31.039815 kernel: NET: Registered PF_INET6 protocol family Sep 13 01:20:31.039821 kernel: Segment Routing with IPv6 Sep 13 01:20:31.039827 kernel: In-situ OAM (IOAM) with IPv6 Sep 13 01:20:31.039833 kernel: NET: Registered PF_PACKET protocol family Sep 13 01:20:31.039839 kernel: Key type dns_resolver registered Sep 13 01:20:31.039845 kernel: microcode: Current revision: 0x00000102 Sep 13 01:20:31.039851 kernel: microcode: Microcode Update Driver: v2.2. Sep 13 01:20:31.039856 kernel: IPI shorthand broadcast: enabled Sep 13 01:20:31.039862 kernel: sched_clock: Marking stable (1904000715, 1379361895)->(4413394452, -1130031842) Sep 13 01:20:31.039868 kernel: registered taskstats version 1 Sep 13 01:20:31.039874 kernel: Loading compiled-in X.509 certificates Sep 13 01:20:31.039880 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.106-flatcar: 1274e0c573ac8d09163d6bc6d1ee1445fb2f8cc6' Sep 13 01:20:31.039885 kernel: Key type .fscrypt registered Sep 13 01:20:31.039892 kernel: Key type fscrypt-provisioning registered Sep 13 01:20:31.039898 kernel: ima: Allocated hash algorithm: sha1 Sep 13 01:20:31.039904 kernel: ima: No architecture policies found Sep 13 01:20:31.039909 kernel: clk: Disabling unused clocks Sep 13 01:20:31.039915 kernel: Freeing unused kernel image (initmem) memory: 42884K Sep 13 01:20:31.039921 kernel: Write protecting the kernel read-only data: 36864k Sep 13 01:20:31.039927 kernel: Freeing unused kernel image (rodata/data gap) memory: 1832K Sep 13 01:20:31.039933 kernel: Run /init as init process Sep 13 01:20:31.039939 kernel: with arguments: Sep 13 01:20:31.039945 kernel: /init Sep 13 01:20:31.039951 kernel: with environment: Sep 13 01:20:31.039957 kernel: HOME=/ Sep 13 01:20:31.039962 kernel: TERM=linux Sep 13 01:20:31.039968 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 13 01:20:31.039975 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 13 01:20:31.039982 systemd[1]: Detected architecture x86-64. Sep 13 01:20:31.039990 systemd[1]: Running in initrd. Sep 13 01:20:31.039996 systemd[1]: No hostname configured, using default hostname. Sep 13 01:20:31.040002 systemd[1]: Hostname set to . Sep 13 01:20:31.040007 systemd[1]: Initializing machine ID from random generator. Sep 13 01:20:31.040014 systemd[1]: Queued start job for default target initrd.target. Sep 13 01:20:31.040020 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:20:31.040026 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:20:31.040032 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 13 01:20:31.040040 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 13 01:20:31.040046 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 13 01:20:31.040052 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 13 01:20:31.040059 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 13 01:20:31.040065 kernel: tsc: Refined TSC clocksource calibration: 3407.999 MHz Sep 13 01:20:31.040071 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd336761, max_idle_ns: 440795243819 ns Sep 13 01:20:31.040077 kernel: clocksource: Switched to clocksource tsc Sep 13 01:20:31.040084 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 13 01:20:31.040090 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:20:31.040096 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:20:31.040102 systemd[1]: Reached target paths.target - Path Units. Sep 13 01:20:31.040109 systemd[1]: Reached target slices.target - Slice Units. Sep 13 01:20:31.040115 systemd[1]: Reached target swap.target - Swaps. Sep 13 01:20:31.040121 systemd[1]: Reached target timers.target - Timer Units. Sep 13 01:20:31.040127 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 13 01:20:31.040134 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 13 01:20:31.040140 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 13 01:20:31.040146 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 13 01:20:31.040152 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:20:31.040158 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 13 01:20:31.040164 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:20:31.040171 systemd[1]: Reached target sockets.target - Socket Units. Sep 13 01:20:31.040177 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 13 01:20:31.040183 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 13 01:20:31.040190 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 13 01:20:31.040196 systemd[1]: Starting systemd-fsck-usr.service... Sep 13 01:20:31.040202 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 13 01:20:31.040219 systemd-journald[265]: Collecting audit messages is disabled. Sep 13 01:20:31.040234 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 13 01:20:31.040241 systemd-journald[265]: Journal started Sep 13 01:20:31.040254 systemd-journald[265]: Runtime Journal (/run/log/journal/376e6e56161f4d3e8ac7e283b728153a) is 8.0M, max 639.9M, 631.9M free. Sep 13 01:20:31.054384 systemd-modules-load[266]: Inserted module 'overlay' Sep 13 01:20:31.076343 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:20:31.104356 systemd[1]: Started systemd-journald.service - Journal Service. Sep 13 01:20:31.104844 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 13 01:20:31.166910 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 13 01:20:31.166926 kernel: Bridge firewalling registered Sep 13 01:20:31.161526 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:20:31.166901 systemd-modules-load[266]: Inserted module 'br_netfilter' Sep 13 01:20:31.186703 systemd[1]: Finished systemd-fsck-usr.service. Sep 13 01:20:31.197746 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 13 01:20:31.226660 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:31.256704 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:20:31.269260 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 13 01:20:31.270872 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 13 01:20:31.272515 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 13 01:20:31.277093 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 13 01:20:31.277793 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 13 01:20:31.277891 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:20:31.278664 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:20:31.279536 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 13 01:20:31.292628 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:31.296957 systemd-resolved[302]: Positive Trust Anchors: Sep 13 01:20:31.296963 systemd-resolved[302]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 01:20:31.296985 systemd-resolved[302]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 13 01:20:31.298591 systemd-resolved[302]: Defaulting to hostname 'linux'. Sep 13 01:20:31.302818 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 13 01:20:31.334687 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:20:31.352718 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:20:31.388720 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 13 01:20:31.502335 dracut-cmdline[307]: dracut-dracut-053 Sep 13 01:20:31.509565 dracut-cmdline[307]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:20:31.703375 kernel: SCSI subsystem initialized Sep 13 01:20:31.726371 kernel: Loading iSCSI transport class v2.0-870. Sep 13 01:20:31.749370 kernel: iscsi: registered transport (tcp) Sep 13 01:20:31.782132 kernel: iscsi: registered transport (qla4xxx) Sep 13 01:20:31.782150 kernel: QLogic iSCSI HBA Driver Sep 13 01:20:31.815193 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 13 01:20:31.835664 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 13 01:20:31.923048 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 13 01:20:31.923068 kernel: device-mapper: uevent: version 1.0.3 Sep 13 01:20:31.942633 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 13 01:20:32.001448 kernel: raid6: avx2x4 gen() 53433 MB/s Sep 13 01:20:32.033375 kernel: raid6: avx2x2 gen() 53370 MB/s Sep 13 01:20:32.069757 kernel: raid6: avx2x1 gen() 45332 MB/s Sep 13 01:20:32.069773 kernel: raid6: using algorithm avx2x4 gen() 53433 MB/s Sep 13 01:20:32.116786 kernel: raid6: .... xor() 17429 MB/s, rmw enabled Sep 13 01:20:32.116806 kernel: raid6: using avx2x2 recovery algorithm Sep 13 01:20:32.157387 kernel: xor: automatically using best checksumming function avx Sep 13 01:20:32.271374 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 13 01:20:32.276849 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 13 01:20:32.310691 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:20:32.317711 systemd-udevd[495]: Using default interface naming scheme 'v255'. Sep 13 01:20:32.321556 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:20:32.355534 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 13 01:20:32.421934 dracut-pre-trigger[506]: rd.md=0: removing MD RAID activation Sep 13 01:20:32.450014 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 13 01:20:32.468580 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 13 01:20:32.542318 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:20:32.587102 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 13 01:20:32.587138 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 13 01:20:32.553316 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 13 01:20:32.590501 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 01:20:32.590532 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:32.603389 kernel: cryptd: max_cpu_qlen set to 1000 Sep 13 01:20:32.613345 kernel: PTP clock support registered Sep 13 01:20:32.613365 kernel: libata version 3.00 loaded. Sep 13 01:20:32.623345 kernel: ACPI: bus type USB registered Sep 13 01:20:32.644785 kernel: usbcore: registered new interface driver usbfs Sep 13 01:20:32.654343 kernel: usbcore: registered new interface driver hub Sep 13 01:20:32.654360 kernel: usbcore: registered new device driver usb Sep 13 01:20:32.683861 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:20:32.771455 kernel: AVX2 version of gcm_enc/dec engaged. Sep 13 01:20:32.771471 kernel: AES CTR mode by8 optimization enabled Sep 13 01:20:32.771479 kernel: ahci 0000:00:17.0: version 3.0 Sep 13 01:20:32.771571 kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 7 ports 6 Gbps 0x7f impl SATA mode Sep 13 01:20:32.771642 kernel: ahci 0000:00:17.0: flags: 64bit ncq sntf clo only pio slum part ems deso sadm sds apst Sep 13 01:20:32.771705 kernel: scsi host0: ahci Sep 13 01:20:32.771772 kernel: scsi host1: ahci Sep 13 01:20:32.731445 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 01:20:32.806659 kernel: scsi host2: ahci Sep 13 01:20:32.806743 kernel: scsi host3: ahci Sep 13 01:20:32.806807 kernel: scsi host4: ahci Sep 13 01:20:32.731493 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:32.980515 kernel: scsi host5: ahci Sep 13 01:20:32.980616 kernel: scsi host6: ahci Sep 13 01:20:32.980696 kernel: ata1: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516100 irq 127 Sep 13 01:20:32.980705 kernel: ata2: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516180 irq 127 Sep 13 01:20:32.980713 kernel: ata3: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516200 irq 127 Sep 13 01:20:32.980720 kernel: ata4: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516280 irq 127 Sep 13 01:20:32.980727 kernel: ata5: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516300 irq 127 Sep 13 01:20:32.980734 kernel: ata6: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516380 irq 127 Sep 13 01:20:32.980741 kernel: ata7: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516400 irq 127 Sep 13 01:20:32.980748 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Sep 13 01:20:32.980756 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Sep 13 01:20:32.794401 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:20:33.089596 kernel: mlx5_core 0000:01:00.0: firmware version: 14.31.1014 Sep 13 01:20:33.089745 kernel: igb 0000:03:00.0: added PHC on eth0 Sep 13 01:20:33.089821 kernel: mlx5_core 0000:01:00.0: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Sep 13 01:20:33.089888 kernel: igb 0000:03:00.0: Intel(R) Gigabit Ethernet Network Connection Sep 13 01:20:33.089958 kernel: igb 0000:03:00.0: eth0: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:32:0c Sep 13 01:20:33.090024 kernel: igb 0000:03:00.0: eth0: PBA No: 010000-000 Sep 13 01:20:33.090088 kernel: igb 0000:03:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Sep 13 01:20:32.980604 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:20:32.996789 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 13 01:20:33.090094 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 13 01:20:33.090119 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:20:33.090146 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 13 01:20:33.099471 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 13 01:20:33.360817 kernel: igb 0000:04:00.0: added PHC on eth1 Sep 13 01:20:33.360913 kernel: igb 0000:04:00.0: Intel(R) Gigabit Ethernet Network Connection Sep 13 01:20:33.360982 kernel: igb 0000:04:00.0: eth1: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:32:0d Sep 13 01:20:33.361048 kernel: igb 0000:04:00.0: eth1: PBA No: 010000-000 Sep 13 01:20:33.361112 kernel: ata2: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Sep 13 01:20:33.361120 kernel: igb 0000:04:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Sep 13 01:20:33.361183 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361194 kernel: ata1: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Sep 13 01:20:33.361201 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361208 kernel: ata3: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361216 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361223 kernel: ata7: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361230 kernel: ata2.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Sep 13 01:20:33.361237 kernel: ata1.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Sep 13 01:20:33.361244 kernel: mlx5_core 0000:01:00.0: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Sep 13 01:20:33.361311 kernel: ata2.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Sep 13 01:20:33.361319 kernel: mlx5_core 0000:01:00.0: Port module event: module 0, Cable plugged Sep 13 01:20:33.361391 kernel: ata1.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Sep 13 01:20:33.361399 kernel: ata1.00: Features: NCQ-prio Sep 13 01:20:33.226628 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:33.383332 kernel: ata2.00: Features: NCQ-prio Sep 13 01:20:33.396353 kernel: ata1.00: configured for UDMA/133 Sep 13 01:20:33.396396 kernel: ata2.00: configured for UDMA/133 Sep 13 01:20:33.401713 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:20:33.441136 kernel: scsi 0:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Sep 13 01:20:33.441289 kernel: scsi 1:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Sep 13 01:20:33.455469 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Sep 13 01:20:33.455587 kernel: igb 0000:03:00.0 eno1: renamed from eth0 Sep 13 01:20:33.455671 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 1 Sep 13 01:20:33.479611 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 13 01:20:33.503380 kernel: xhci_hcd 0000:00:14.0: hcc params 0x200077c1 hci version 0x110 quirks 0x0000000000009810 Sep 13 01:20:33.515441 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Sep 13 01:20:33.515564 kernel: igb 0000:04:00.0 eno2: renamed from eth1 Sep 13 01:20:33.515654 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 2 Sep 13 01:20:33.555348 kernel: xhci_hcd 0000:00:14.0: Host supports USB 3.1 Enhanced SuperSpeed Sep 13 01:20:33.564344 kernel: hub 1-0:1.0: USB hub found Sep 13 01:20:33.579976 kernel: hub 1-0:1.0: 16 ports detected Sep 13 01:20:33.602343 kernel: hub 2-0:1.0: USB hub found Sep 13 01:20:33.602436 kernel: mlx5_core 0000:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Sep 13 01:20:33.602511 kernel: hub 2-0:1.0: 10 ports detected Sep 13 01:20:33.615345 kernel: mlx5_core 0000:01:00.1: firmware version: 14.31.1014 Sep 13 01:20:33.656304 kernel: mlx5_core 0000:01:00.1: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Sep 13 01:20:33.663068 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:33.794297 kernel: ata2.00: Enabling discard_zeroes_data Sep 13 01:20:33.794311 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:33.794319 kernel: sd 1:0:0:0: [sda] 937703088 512-byte logical blocks: (480 GB/447 GiB) Sep 13 01:20:33.794409 kernel: sd 0:0:0:0: [sdb] 937703088 512-byte logical blocks: (480 GB/447 GiB) Sep 13 01:20:33.794478 kernel: sd 1:0:0:0: [sda] 4096-byte physical blocks Sep 13 01:20:33.794540 kernel: sd 0:0:0:0: [sdb] 4096-byte physical blocks Sep 13 01:20:33.794603 kernel: sd 1:0:0:0: [sda] Write Protect is off Sep 13 01:20:33.794665 kernel: sd 1:0:0:0: [sda] Mode Sense: 00 3a 00 00 Sep 13 01:20:33.794727 kernel: sd 1:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Sep 13 01:20:33.794788 kernel: sd 1:0:0:0: [sda] Preferred minimum I/O size 4096 bytes Sep 13 01:20:33.794849 kernel: sd 0:0:0:0: [sdb] Write Protect is off Sep 13 01:20:33.794908 kernel: ata2.00: Enabling discard_zeroes_data Sep 13 01:20:33.794916 kernel: sd 1:0:0:0: [sda] Attached SCSI disk Sep 13 01:20:33.840846 kernel: usb 1-14: new high-speed USB device number 2 using xhci_hcd Sep 13 01:20:33.840876 kernel: sd 0:0:0:0: [sdb] Mode Sense: 00 3a 00 00 Sep 13 01:20:33.876005 kernel: sd 0:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Sep 13 01:20:33.885399 kernel: sd 0:0:0:0: [sdb] Preferred minimum I/O size 4096 bytes Sep 13 01:20:33.900409 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:33.935438 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 13 01:20:33.935456 kernel: GPT:9289727 != 937703087 Sep 13 01:20:33.950220 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 13 01:20:33.962829 kernel: GPT:9289727 != 937703087 Sep 13 01:20:33.962848 kernel: mlx5_core 0000:01:00.1: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Sep 13 01:20:33.962935 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 13 01:20:33.962944 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:33.967379 kernel: hub 1-14:1.0: USB hub found Sep 13 01:20:33.967473 kernel: hub 1-14:1.0: 4 ports detected Sep 13 01:20:33.990980 kernel: mlx5_core 0000:01:00.1: Port module event: module 1, Cable plugged Sep 13 01:20:33.991069 kernel: sd 0:0:0:0: [sdb] Attached SCSI disk Sep 13 01:20:34.110459 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Micron_5300_MTFDDAK480TDT ROOT. Sep 13 01:20:34.165588 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sdb6 scanned by (udev-worker) (680) Sep 13 01:20:34.165603 kernel: BTRFS: device fsid fa70a3b0-3d47-4508-bba0-9fa4607626aa devid 1 transid 36 /dev/sdb3 scanned by (udev-worker) (544) Sep 13 01:20:34.149084 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Micron_5300_MTFDDAK480TDT EFI-SYSTEM. Sep 13 01:20:34.184632 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Sep 13 01:20:34.192981 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Micron_5300_MTFDDAK480TDT USR-A. Sep 13 01:20:34.221605 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Micron_5300_MTFDDAK480TDT USR-A. Sep 13 01:20:34.306582 kernel: mlx5_core 0000:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Sep 13 01:20:34.306678 kernel: usb 1-14.1: new low-speed USB device number 3 using xhci_hcd Sep 13 01:20:34.306697 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: renamed from eth0 Sep 13 01:20:34.307515 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 13 01:20:34.358394 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: renamed from eth1 Sep 13 01:20:34.358486 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:34.358495 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:34.358539 disk-uuid[731]: Primary Header is updated. Sep 13 01:20:34.358539 disk-uuid[731]: Secondary Entries is updated. Sep 13 01:20:34.358539 disk-uuid[731]: Secondary Header is updated. Sep 13 01:20:34.407454 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:34.407466 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:34.407473 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:34.430345 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:34.453405 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 13 01:20:34.475584 kernel: usbcore: registered new interface driver usbhid Sep 13 01:20:34.475612 kernel: usbhid: USB HID core driver Sep 13 01:20:34.520414 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.0/0003:0557:2419.0001/input/input0 Sep 13 01:20:34.616423 kernel: hid-generic 0003:0557:2419.0001: input,hidraw0: USB HID v1.00 Keyboard [HID 0557:2419] on usb-0000:00:14.0-14.1/input0 Sep 13 01:20:34.616546 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.1/0003:0557:2419.0002/input/input1 Sep 13 01:20:34.650314 kernel: hid-generic 0003:0557:2419.0002: input,hidraw1: USB HID v1.00 Mouse [HID 0557:2419] on usb-0000:00:14.0-14.1/input1 Sep 13 01:20:35.413889 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:35.433278 disk-uuid[733]: The operation has completed successfully. Sep 13 01:20:35.441450 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:35.475290 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 13 01:20:35.475353 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 13 01:20:35.505630 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 13 01:20:35.532133 sh[752]: Success Sep 13 01:20:35.560383 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 13 01:20:35.610995 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 13 01:20:35.641591 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 13 01:20:35.649656 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 13 01:20:35.705201 kernel: BTRFS info (device dm-0): first mount of filesystem fa70a3b0-3d47-4508-bba0-9fa4607626aa Sep 13 01:20:35.705226 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:35.727003 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 13 01:20:35.746501 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 13 01:20:35.764911 kernel: BTRFS info (device dm-0): using free space tree Sep 13 01:20:35.805375 kernel: BTRFS info (device dm-0): enabling ssd optimizations Sep 13 01:20:35.808108 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 13 01:20:35.816778 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 13 01:20:35.830598 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 13 01:20:35.848752 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 13 01:20:35.892891 kernel: BTRFS info (device sdb6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:35.892912 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:35.911481 kernel: BTRFS info (device sdb6): using free space tree Sep 13 01:20:35.937604 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 13 01:20:35.994621 kernel: BTRFS info (device sdb6): enabling ssd optimizations Sep 13 01:20:35.994638 kernel: BTRFS info (device sdb6): auto enabling async discard Sep 13 01:20:35.994646 kernel: BTRFS info (device sdb6): last unmount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:35.997638 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 13 01:20:36.007627 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 13 01:20:36.008649 systemd-networkd[933]: lo: Link UP Sep 13 01:20:36.008651 systemd-networkd[933]: lo: Gained carrier Sep 13 01:20:36.011122 systemd-networkd[933]: Enumeration completed Sep 13 01:20:36.011866 systemd-networkd[933]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.017684 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 13 01:20:36.041456 systemd-networkd[933]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.044872 systemd[1]: Reached target network.target - Network. Sep 13 01:20:36.069921 systemd-networkd[933]: enp1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.073499 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 13 01:20:36.143864 unknown[937]: fetched base config from "system" Sep 13 01:20:36.141750 ignition[937]: Ignition 2.19.0 Sep 13 01:20:36.143869 unknown[937]: fetched user config from "system" Sep 13 01:20:36.141754 ignition[937]: Stage: fetch-offline Sep 13 01:20:36.145155 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 13 01:20:36.141778 ignition[937]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:36.158652 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 13 01:20:36.141784 ignition[937]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:36.172701 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 13 01:20:36.266593 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Sep 13 01:20:36.141841 ignition[937]: parsed url from cmdline: "" Sep 13 01:20:36.260800 systemd-networkd[933]: enp1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.141843 ignition[937]: no config URL provided Sep 13 01:20:36.141846 ignition[937]: reading system config file "/usr/lib/ignition/user.ign" Sep 13 01:20:36.141872 ignition[937]: parsing config with SHA512: 26b2c5bf2a30dc9da010c8bdf3f2719492292f4d9470ffc922d9935718c9a83cfef36d87e1f121dc535c90050c9d5379081ae80fca0014dc6a76082ef4a3c8c8 Sep 13 01:20:36.144072 ignition[937]: fetch-offline: fetch-offline passed Sep 13 01:20:36.144075 ignition[937]: POST message to Packet Timeline Sep 13 01:20:36.144077 ignition[937]: POST Status error: resource requires networking Sep 13 01:20:36.144364 ignition[937]: Ignition finished successfully Sep 13 01:20:36.183536 ignition[950]: Ignition 2.19.0 Sep 13 01:20:36.183543 ignition[950]: Stage: kargs Sep 13 01:20:36.183709 ignition[950]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:36.183720 ignition[950]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:36.184596 ignition[950]: kargs: kargs passed Sep 13 01:20:36.184601 ignition[950]: POST message to Packet Timeline Sep 13 01:20:36.184614 ignition[950]: GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:36.185273 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:43970->[::1]:53: read: connection refused Sep 13 01:20:36.385986 ignition[950]: GET https://metadata.packet.net/metadata: attempt #2 Sep 13 01:20:36.387247 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:58372->[::1]:53: read: connection refused Sep 13 01:20:36.520447 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Sep 13 01:20:36.523332 systemd-networkd[933]: eno1: Link UP Sep 13 01:20:36.523478 systemd-networkd[933]: eno2: Link UP Sep 13 01:20:36.523601 systemd-networkd[933]: enp1s0f0np0: Link UP Sep 13 01:20:36.523742 systemd-networkd[933]: enp1s0f0np0: Gained carrier Sep 13 01:20:36.534534 systemd-networkd[933]: enp1s0f1np1: Link UP Sep 13 01:20:36.559448 systemd-networkd[933]: enp1s0f0np0: DHCPv4 address 139.178.94.15/31, gateway 139.178.94.14 acquired from 145.40.83.140 Sep 13 01:20:36.787524 ignition[950]: GET https://metadata.packet.net/metadata: attempt #3 Sep 13 01:20:36.788670 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:48951->[::1]:53: read: connection refused Sep 13 01:20:37.308873 systemd-networkd[933]: enp1s0f1np1: Gained carrier Sep 13 01:20:37.588994 ignition[950]: GET https://metadata.packet.net/metadata: attempt #4 Sep 13 01:20:37.590003 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:57512->[::1]:53: read: connection refused Sep 13 01:20:38.012945 systemd-networkd[933]: enp1s0f0np0: Gained IPv6LL Sep 13 01:20:38.396671 systemd-networkd[933]: enp1s0f1np1: Gained IPv6LL Sep 13 01:20:39.190579 ignition[950]: GET https://metadata.packet.net/metadata: attempt #5 Sep 13 01:20:39.191662 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:40537->[::1]:53: read: connection refused Sep 13 01:20:42.394328 ignition[950]: GET https://metadata.packet.net/metadata: attempt #6 Sep 13 01:20:43.402394 ignition[950]: GET result: OK Sep 13 01:20:43.806538 ignition[950]: Ignition finished successfully Sep 13 01:20:43.810824 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 13 01:20:43.834618 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 13 01:20:43.840733 ignition[969]: Ignition 2.19.0 Sep 13 01:20:43.840738 ignition[969]: Stage: disks Sep 13 01:20:43.840842 ignition[969]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:43.840848 ignition[969]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:43.841362 ignition[969]: disks: disks passed Sep 13 01:20:43.841365 ignition[969]: POST message to Packet Timeline Sep 13 01:20:43.841374 ignition[969]: GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:45.035326 ignition[969]: GET result: OK Sep 13 01:20:45.509175 ignition[969]: Ignition finished successfully Sep 13 01:20:45.512466 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 13 01:20:45.527583 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 13 01:20:45.545605 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 13 01:20:45.556073 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 13 01:20:45.587644 systemd[1]: Reached target sysinit.target - System Initialization. Sep 13 01:20:45.607738 systemd[1]: Reached target basic.target - Basic System. Sep 13 01:20:45.640586 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 13 01:20:45.674920 systemd-fsck[987]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 13 01:20:45.686132 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 13 01:20:45.705676 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 13 01:20:45.808406 kernel: EXT4-fs (sdb9): mounted filesystem 3a3ecd49-b269-4fcb-bb61-e2994e1868ee r/w with ordered data mode. Quota mode: none. Sep 13 01:20:45.808887 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 13 01:20:45.818784 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 13 01:20:45.855563 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 13 01:20:45.864494 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 13 01:20:45.989727 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sdb6 scanned by mount (997) Sep 13 01:20:45.989741 kernel: BTRFS info (device sdb6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:45.989749 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:45.989756 kernel: BTRFS info (device sdb6): using free space tree Sep 13 01:20:45.989763 kernel: BTRFS info (device sdb6): enabling ssd optimizations Sep 13 01:20:45.989770 kernel: BTRFS info (device sdb6): auto enabling async discard Sep 13 01:20:45.904984 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Sep 13 01:20:45.990156 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Sep 13 01:20:46.018603 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 13 01:20:46.018621 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 13 01:20:46.085443 coreos-metadata[999]: Sep 13 01:20:46.065 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Sep 13 01:20:46.038549 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 13 01:20:46.106649 coreos-metadata[1015]: Sep 13 01:20:46.070 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Sep 13 01:20:46.076540 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 13 01:20:46.106578 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 13 01:20:46.152514 initrd-setup-root[1029]: cut: /sysroot/etc/passwd: No such file or directory Sep 13 01:20:46.162589 initrd-setup-root[1036]: cut: /sysroot/etc/group: No such file or directory Sep 13 01:20:46.172557 initrd-setup-root[1043]: cut: /sysroot/etc/shadow: No such file or directory Sep 13 01:20:46.182464 initrd-setup-root[1050]: cut: /sysroot/etc/gshadow: No such file or directory Sep 13 01:20:46.202365 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 13 01:20:46.222579 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 13 01:20:46.258582 kernel: BTRFS info (device sdb6): last unmount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:46.241872 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 13 01:20:46.259145 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 13 01:20:46.289838 ignition[1117]: INFO : Ignition 2.19.0 Sep 13 01:20:46.289838 ignition[1117]: INFO : Stage: mount Sep 13 01:20:46.304450 ignition[1117]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:46.304450 ignition[1117]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:46.304450 ignition[1117]: INFO : mount: mount passed Sep 13 01:20:46.304450 ignition[1117]: INFO : POST message to Packet Timeline Sep 13 01:20:46.304450 ignition[1117]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:46.302085 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 13 01:20:47.161379 coreos-metadata[1015]: Sep 13 01:20:47.161 INFO Fetch successful Sep 13 01:20:47.243448 systemd[1]: flatcar-static-network.service: Deactivated successfully. Sep 13 01:20:47.243506 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Sep 13 01:20:47.276426 coreos-metadata[999]: Sep 13 01:20:47.260 INFO Fetch successful Sep 13 01:20:47.288874 coreos-metadata[999]: Sep 13 01:20:47.288 INFO wrote hostname ci-4081.3.5-n-7cd54b3887 to /sysroot/etc/hostname Sep 13 01:20:47.290007 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 13 01:20:47.353546 ignition[1117]: INFO : GET result: OK Sep 13 01:20:48.080554 ignition[1117]: INFO : Ignition finished successfully Sep 13 01:20:48.083719 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 13 01:20:48.115693 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 13 01:20:48.119827 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 13 01:20:48.186344 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sdb6 scanned by mount (1142) Sep 13 01:20:48.215867 kernel: BTRFS info (device sdb6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:48.215883 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:48.233877 kernel: BTRFS info (device sdb6): using free space tree Sep 13 01:20:48.273556 kernel: BTRFS info (device sdb6): enabling ssd optimizations Sep 13 01:20:48.273572 kernel: BTRFS info (device sdb6): auto enabling async discard Sep 13 01:20:48.288098 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 13 01:20:48.315680 ignition[1159]: INFO : Ignition 2.19.0 Sep 13 01:20:48.315680 ignition[1159]: INFO : Stage: files Sep 13 01:20:48.330584 ignition[1159]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:48.330584 ignition[1159]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:48.330584 ignition[1159]: DEBUG : files: compiled without relabeling support, skipping Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 13 01:20:48.330584 ignition[1159]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Sep 13 01:20:48.319288 unknown[1159]: wrote ssh authorized keys file for user: core Sep 13 01:20:48.466585 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 13 01:20:48.575905 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 13 01:20:48.575905 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Sep 13 01:20:49.141995 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 13 01:20:49.782539 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:49.782539 ignition[1159]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: files passed Sep 13 01:20:49.811645 ignition[1159]: INFO : POST message to Packet Timeline Sep 13 01:20:49.811645 ignition[1159]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:50.704146 ignition[1159]: INFO : GET result: OK Sep 13 01:20:51.142646 ignition[1159]: INFO : Ignition finished successfully Sep 13 01:20:51.146583 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 13 01:20:51.184826 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 13 01:20:51.195067 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 13 01:20:51.215726 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 13 01:20:51.215806 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 13 01:20:51.267612 initrd-setup-root-after-ignition[1195]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:20:51.267612 initrd-setup-root-after-ignition[1195]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:20:51.239118 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 13 01:20:51.327547 initrd-setup-root-after-ignition[1199]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:20:51.259725 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 13 01:20:51.292644 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 13 01:20:51.366946 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 13 01:20:51.367195 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 13 01:20:51.387508 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 13 01:20:51.407605 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 13 01:20:51.428933 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 13 01:20:51.442744 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 13 01:20:51.517586 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 13 01:20:51.545810 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 13 01:20:51.564237 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:20:51.568536 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:20:51.600762 systemd[1]: Stopped target timers.target - Timer Units. Sep 13 01:20:51.619667 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 13 01:20:51.619822 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 13 01:20:51.647039 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 13 01:20:51.669065 systemd[1]: Stopped target basic.target - Basic System. Sep 13 01:20:51.688071 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 13 01:20:51.707080 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 13 01:20:51.727954 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 13 01:20:51.748974 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 13 01:20:51.769068 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 13 01:20:51.789988 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 13 01:20:51.811094 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 13 01:20:51.832059 systemd[1]: Stopped target swap.target - Swaps. Sep 13 01:20:51.850828 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 13 01:20:51.851229 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 13 01:20:51.877198 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:20:51.896982 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:20:51.917827 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 13 01:20:51.918289 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:20:51.940962 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 13 01:20:51.941391 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 13 01:20:51.972927 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 13 01:20:51.973394 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 13 01:20:51.993159 systemd[1]: Stopped target paths.target - Path Units. Sep 13 01:20:52.011813 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 13 01:20:52.012289 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:20:52.032962 systemd[1]: Stopped target slices.target - Slice Units. Sep 13 01:20:52.053066 systemd[1]: Stopped target sockets.target - Socket Units. Sep 13 01:20:52.071916 systemd[1]: iscsid.socket: Deactivated successfully. Sep 13 01:20:52.072219 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 13 01:20:52.093110 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 13 01:20:52.093447 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 13 01:20:52.116152 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 13 01:20:52.116585 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 13 01:20:52.136044 systemd[1]: ignition-files.service: Deactivated successfully. Sep 13 01:20:52.248756 ignition[1221]: INFO : Ignition 2.19.0 Sep 13 01:20:52.248756 ignition[1221]: INFO : Stage: umount Sep 13 01:20:52.248756 ignition[1221]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:52.248756 ignition[1221]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:52.248756 ignition[1221]: INFO : umount: umount passed Sep 13 01:20:52.248756 ignition[1221]: INFO : POST message to Packet Timeline Sep 13 01:20:52.248756 ignition[1221]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:52.136447 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 13 01:20:52.154030 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Sep 13 01:20:52.154446 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 13 01:20:52.182620 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 13 01:20:52.200621 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 13 01:20:52.200741 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:20:52.239708 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 13 01:20:52.240644 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 13 01:20:52.240859 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:20:52.266919 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 13 01:20:52.267276 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 13 01:20:52.306581 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 13 01:20:52.306926 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 13 01:20:52.306970 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 13 01:20:52.329618 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 13 01:20:52.329682 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 13 01:20:53.291066 ignition[1221]: INFO : GET result: OK Sep 13 01:20:54.076937 ignition[1221]: INFO : Ignition finished successfully Sep 13 01:20:54.080070 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 13 01:20:54.080391 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 13 01:20:54.096649 systemd[1]: Stopped target network.target - Network. Sep 13 01:20:54.112581 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 13 01:20:54.112750 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 13 01:20:54.130670 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 13 01:20:54.130818 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 13 01:20:54.148637 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 13 01:20:54.148760 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 13 01:20:54.166866 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 13 01:20:54.167038 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 13 01:20:54.185841 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 13 01:20:54.186036 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 13 01:20:54.205256 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 13 01:20:54.215497 systemd-networkd[933]: enp1s0f0np0: DHCPv6 lease lost Sep 13 01:20:54.224757 systemd-networkd[933]: enp1s0f1np1: DHCPv6 lease lost Sep 13 01:20:54.225125 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 13 01:20:54.245465 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 13 01:20:54.245757 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 13 01:20:54.265524 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 13 01:20:54.265823 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 13 01:20:54.287241 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 13 01:20:54.287474 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:20:54.317633 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 13 01:20:54.325701 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 13 01:20:54.325845 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 13 01:20:54.354666 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 01:20:54.354798 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:20:54.372734 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 13 01:20:54.372891 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 13 01:20:54.390758 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 13 01:20:54.390926 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:20:54.411980 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:20:54.432648 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 13 01:20:54.433031 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:20:54.463751 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 13 01:20:54.463788 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 13 01:20:54.471695 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 13 01:20:54.471718 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:20:54.499642 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 13 01:20:54.499712 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 13 01:20:54.540554 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 13 01:20:54.540802 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 13 01:20:54.579465 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 01:20:54.579706 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:54.622437 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 13 01:20:54.650406 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 13 01:20:54.865596 systemd-journald[265]: Received SIGTERM from PID 1 (systemd). Sep 13 01:20:54.650448 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:20:54.669489 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 01:20:54.669571 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:54.692696 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 13 01:20:54.692938 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 13 01:20:54.724172 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 13 01:20:54.724453 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 13 01:20:54.744606 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 13 01:20:54.780757 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 13 01:20:54.801433 systemd[1]: Switching root. Sep 13 01:20:54.959527 systemd-journald[265]: Journal stopped Sep 13 01:20:31.026584 kernel: Linux version 6.6.106-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Sep 12 22:30:50 -00 2025 Sep 13 01:20:31.026598 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:20:31.026605 kernel: BIOS-provided physical RAM map: Sep 13 01:20:31.026609 kernel: BIOS-e820: [mem 0x0000000000000000-0x00000000000997ff] usable Sep 13 01:20:31.026613 kernel: BIOS-e820: [mem 0x0000000000099800-0x000000000009ffff] reserved Sep 13 01:20:31.026617 kernel: BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved Sep 13 01:20:31.026622 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003fffffff] usable Sep 13 01:20:31.026626 kernel: BIOS-e820: [mem 0x0000000040000000-0x00000000403fffff] reserved Sep 13 01:20:31.026630 kernel: BIOS-e820: [mem 0x0000000040400000-0x00000000825bdfff] usable Sep 13 01:20:31.026635 kernel: BIOS-e820: [mem 0x00000000825be000-0x00000000825befff] ACPI NVS Sep 13 01:20:31.026639 kernel: BIOS-e820: [mem 0x00000000825bf000-0x00000000825bffff] reserved Sep 13 01:20:31.026644 kernel: BIOS-e820: [mem 0x00000000825c0000-0x000000008afcdfff] usable Sep 13 01:20:31.026648 kernel: BIOS-e820: [mem 0x000000008afce000-0x000000008c0b2fff] reserved Sep 13 01:20:31.026653 kernel: BIOS-e820: [mem 0x000000008c0b3000-0x000000008c23bfff] usable Sep 13 01:20:31.026658 kernel: BIOS-e820: [mem 0x000000008c23c000-0x000000008c66dfff] ACPI NVS Sep 13 01:20:31.026663 kernel: BIOS-e820: [mem 0x000000008c66e000-0x000000008eefefff] reserved Sep 13 01:20:31.026668 kernel: BIOS-e820: [mem 0x000000008eeff000-0x000000008eefffff] usable Sep 13 01:20:31.026673 kernel: BIOS-e820: [mem 0x000000008ef00000-0x000000008fffffff] reserved Sep 13 01:20:31.026678 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 13 01:20:31.026682 kernel: BIOS-e820: [mem 0x00000000fe000000-0x00000000fe010fff] reserved Sep 13 01:20:31.026687 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec00fff] reserved Sep 13 01:20:31.026692 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Sep 13 01:20:31.026696 kernel: BIOS-e820: [mem 0x00000000ff000000-0x00000000ffffffff] reserved Sep 13 01:20:31.026701 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000086effffff] usable Sep 13 01:20:31.026706 kernel: NX (Execute Disable) protection: active Sep 13 01:20:31.026711 kernel: APIC: Static calls initialized Sep 13 01:20:31.026715 kernel: SMBIOS 3.2.1 present. Sep 13 01:20:31.026720 kernel: DMI: Supermicro SYS-5019C-MR-PH004/X11SCM-F, BIOS 2.6 12/03/2024 Sep 13 01:20:31.026726 kernel: tsc: Detected 3400.000 MHz processor Sep 13 01:20:31.026731 kernel: tsc: Detected 3399.906 MHz TSC Sep 13 01:20:31.026736 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 13 01:20:31.026741 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 13 01:20:31.026746 kernel: last_pfn = 0x86f000 max_arch_pfn = 0x400000000 Sep 13 01:20:31.026751 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 23), built from 10 variable MTRRs Sep 13 01:20:31.026756 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 13 01:20:31.026760 kernel: last_pfn = 0x8ef00 max_arch_pfn = 0x400000000 Sep 13 01:20:31.026765 kernel: Using GB pages for direct mapping Sep 13 01:20:31.026771 kernel: ACPI: Early table checksum verification disabled Sep 13 01:20:31.026776 kernel: ACPI: RSDP 0x00000000000F05B0 000024 (v02 SUPERM) Sep 13 01:20:31.026781 kernel: ACPI: XSDT 0x000000008C54F0C8 00010C (v01 SUPERM SUPERM 01072009 AMI 00010013) Sep 13 01:20:31.026788 kernel: ACPI: FACP 0x000000008C58B670 000114 (v06 01072009 AMI 00010013) Sep 13 01:20:31.026793 kernel: ACPI: DSDT 0x000000008C54F268 03C404 (v02 SUPERM SMCI--MB 01072009 INTL 20160527) Sep 13 01:20:31.026798 kernel: ACPI: FACS 0x000000008C66DF80 000040 Sep 13 01:20:31.026803 kernel: ACPI: APIC 0x000000008C58B788 00012C (v04 01072009 AMI 00010013) Sep 13 01:20:31.026810 kernel: ACPI: FPDT 0x000000008C58B8B8 000044 (v01 01072009 AMI 00010013) Sep 13 01:20:31.026815 kernel: ACPI: FIDT 0x000000008C58B900 00009C (v01 SUPERM SMCI--MB 01072009 AMI 00010013) Sep 13 01:20:31.026820 kernel: ACPI: MCFG 0x000000008C58B9A0 00003C (v01 SUPERM SMCI--MB 01072009 MSFT 00000097) Sep 13 01:20:31.026825 kernel: ACPI: SPMI 0x000000008C58B9E0 000041 (v05 SUPERM SMCI--MB 00000000 AMI. 00000000) Sep 13 01:20:31.026830 kernel: ACPI: SSDT 0x000000008C58BA28 001B1C (v02 CpuRef CpuSsdt 00003000 INTL 20160527) Sep 13 01:20:31.026835 kernel: ACPI: SSDT 0x000000008C58D548 0031C6 (v02 SaSsdt SaSsdt 00003000 INTL 20160527) Sep 13 01:20:31.026840 kernel: ACPI: SSDT 0x000000008C590710 00232B (v02 PegSsd PegSsdt 00001000 INTL 20160527) Sep 13 01:20:31.026846 kernel: ACPI: HPET 0x000000008C592A40 000038 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026852 kernel: ACPI: SSDT 0x000000008C592A78 000FAE (v02 SUPERM Ther_Rvp 00001000 INTL 20160527) Sep 13 01:20:31.026857 kernel: ACPI: SSDT 0x000000008C593A28 0008F4 (v02 INTEL xh_mossb 00000000 INTL 20160527) Sep 13 01:20:31.026862 kernel: ACPI: UEFI 0x000000008C594320 000042 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026867 kernel: ACPI: LPIT 0x000000008C594368 000094 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026872 kernel: ACPI: SSDT 0x000000008C594400 0027DE (v02 SUPERM PtidDevc 00001000 INTL 20160527) Sep 13 01:20:31.026877 kernel: ACPI: SSDT 0x000000008C596BE0 0014E2 (v02 SUPERM TbtTypeC 00000000 INTL 20160527) Sep 13 01:20:31.026882 kernel: ACPI: DBGP 0x000000008C5980C8 000034 (v01 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026887 kernel: ACPI: DBG2 0x000000008C598100 000054 (v00 SUPERM SMCI--MB 00000002 01000013) Sep 13 01:20:31.026893 kernel: ACPI: SSDT 0x000000008C598158 001B67 (v02 SUPERM UsbCTabl 00001000 INTL 20160527) Sep 13 01:20:31.026899 kernel: ACPI: DMAR 0x000000008C599CC0 000070 (v01 INTEL EDK2 00000002 01000013) Sep 13 01:20:31.026904 kernel: ACPI: SSDT 0x000000008C599D30 000144 (v02 Intel ADebTabl 00001000 INTL 20160527) Sep 13 01:20:31.026909 kernel: ACPI: TPM2 0x000000008C599E78 000034 (v04 SUPERM SMCI--MB 00000001 AMI 00000000) Sep 13 01:20:31.026914 kernel: ACPI: SSDT 0x000000008C599EB0 000D8F (v02 INTEL SpsNm 00000002 INTL 20160527) Sep 13 01:20:31.026919 kernel: ACPI: WSMT 0x000000008C59AC40 000028 (v01 SUPERM 01072009 AMI 00010013) Sep 13 01:20:31.026925 kernel: ACPI: EINJ 0x000000008C59AC68 000130 (v01 AMI AMI.EINJ 00000000 AMI. 00000000) Sep 13 01:20:31.026930 kernel: ACPI: ERST 0x000000008C59AD98 000230 (v01 AMIER AMI.ERST 00000000 AMI. 00000000) Sep 13 01:20:31.026936 kernel: ACPI: BERT 0x000000008C59AFC8 000030 (v01 AMI AMI.BERT 00000000 AMI. 00000000) Sep 13 01:20:31.026941 kernel: ACPI: HEST 0x000000008C59AFF8 00027C (v01 AMI AMI.HEST 00000000 AMI. 00000000) Sep 13 01:20:31.026946 kernel: ACPI: SSDT 0x000000008C59B278 000162 (v01 SUPERM SMCCDN 00000000 INTL 20181221) Sep 13 01:20:31.026951 kernel: ACPI: Reserving FACP table memory at [mem 0x8c58b670-0x8c58b783] Sep 13 01:20:31.026956 kernel: ACPI: Reserving DSDT table memory at [mem 0x8c54f268-0x8c58b66b] Sep 13 01:20:31.026961 kernel: ACPI: Reserving FACS table memory at [mem 0x8c66df80-0x8c66dfbf] Sep 13 01:20:31.026966 kernel: ACPI: Reserving APIC table memory at [mem 0x8c58b788-0x8c58b8b3] Sep 13 01:20:31.026971 kernel: ACPI: Reserving FPDT table memory at [mem 0x8c58b8b8-0x8c58b8fb] Sep 13 01:20:31.026977 kernel: ACPI: Reserving FIDT table memory at [mem 0x8c58b900-0x8c58b99b] Sep 13 01:20:31.026983 kernel: ACPI: Reserving MCFG table memory at [mem 0x8c58b9a0-0x8c58b9db] Sep 13 01:20:31.026988 kernel: ACPI: Reserving SPMI table memory at [mem 0x8c58b9e0-0x8c58ba20] Sep 13 01:20:31.026993 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58ba28-0x8c58d543] Sep 13 01:20:31.026998 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58d548-0x8c59070d] Sep 13 01:20:31.027003 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c590710-0x8c592a3a] Sep 13 01:20:31.027008 kernel: ACPI: Reserving HPET table memory at [mem 0x8c592a40-0x8c592a77] Sep 13 01:20:31.027013 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c592a78-0x8c593a25] Sep 13 01:20:31.027018 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c593a28-0x8c59431b] Sep 13 01:20:31.027023 kernel: ACPI: Reserving UEFI table memory at [mem 0x8c594320-0x8c594361] Sep 13 01:20:31.027029 kernel: ACPI: Reserving LPIT table memory at [mem 0x8c594368-0x8c5943fb] Sep 13 01:20:31.027035 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c594400-0x8c596bdd] Sep 13 01:20:31.027040 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c596be0-0x8c5980c1] Sep 13 01:20:31.027045 kernel: ACPI: Reserving DBGP table memory at [mem 0x8c5980c8-0x8c5980fb] Sep 13 01:20:31.027050 kernel: ACPI: Reserving DBG2 table memory at [mem 0x8c598100-0x8c598153] Sep 13 01:20:31.027055 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c598158-0x8c599cbe] Sep 13 01:20:31.027060 kernel: ACPI: Reserving DMAR table memory at [mem 0x8c599cc0-0x8c599d2f] Sep 13 01:20:31.027065 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599d30-0x8c599e73] Sep 13 01:20:31.027070 kernel: ACPI: Reserving TPM2 table memory at [mem 0x8c599e78-0x8c599eab] Sep 13 01:20:31.027076 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599eb0-0x8c59ac3e] Sep 13 01:20:31.027081 kernel: ACPI: Reserving WSMT table memory at [mem 0x8c59ac40-0x8c59ac67] Sep 13 01:20:31.027086 kernel: ACPI: Reserving EINJ table memory at [mem 0x8c59ac68-0x8c59ad97] Sep 13 01:20:31.027092 kernel: ACPI: Reserving ERST table memory at [mem 0x8c59ad98-0x8c59afc7] Sep 13 01:20:31.027097 kernel: ACPI: Reserving BERT table memory at [mem 0x8c59afc8-0x8c59aff7] Sep 13 01:20:31.027102 kernel: ACPI: Reserving HEST table memory at [mem 0x8c59aff8-0x8c59b273] Sep 13 01:20:31.027107 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c59b278-0x8c59b3d9] Sep 13 01:20:31.027112 kernel: No NUMA configuration found Sep 13 01:20:31.027117 kernel: Faking a node at [mem 0x0000000000000000-0x000000086effffff] Sep 13 01:20:31.027122 kernel: NODE_DATA(0) allocated [mem 0x86effa000-0x86effffff] Sep 13 01:20:31.027129 kernel: Zone ranges: Sep 13 01:20:31.027134 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 13 01:20:31.027139 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Sep 13 01:20:31.027144 kernel: Normal [mem 0x0000000100000000-0x000000086effffff] Sep 13 01:20:31.027149 kernel: Movable zone start for each node Sep 13 01:20:31.027154 kernel: Early memory node ranges Sep 13 01:20:31.027159 kernel: node 0: [mem 0x0000000000001000-0x0000000000098fff] Sep 13 01:20:31.027164 kernel: node 0: [mem 0x0000000000100000-0x000000003fffffff] Sep 13 01:20:31.027170 kernel: node 0: [mem 0x0000000040400000-0x00000000825bdfff] Sep 13 01:20:31.027176 kernel: node 0: [mem 0x00000000825c0000-0x000000008afcdfff] Sep 13 01:20:31.027181 kernel: node 0: [mem 0x000000008c0b3000-0x000000008c23bfff] Sep 13 01:20:31.027186 kernel: node 0: [mem 0x000000008eeff000-0x000000008eefffff] Sep 13 01:20:31.027191 kernel: node 0: [mem 0x0000000100000000-0x000000086effffff] Sep 13 01:20:31.027200 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000086effffff] Sep 13 01:20:31.027206 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 13 01:20:31.027211 kernel: On node 0, zone DMA: 103 pages in unavailable ranges Sep 13 01:20:31.027217 kernel: On node 0, zone DMA32: 1024 pages in unavailable ranges Sep 13 01:20:31.027223 kernel: On node 0, zone DMA32: 2 pages in unavailable ranges Sep 13 01:20:31.027228 kernel: On node 0, zone DMA32: 4325 pages in unavailable ranges Sep 13 01:20:31.027234 kernel: On node 0, zone DMA32: 11459 pages in unavailable ranges Sep 13 01:20:31.027239 kernel: On node 0, zone Normal: 4352 pages in unavailable ranges Sep 13 01:20:31.027245 kernel: On node 0, zone Normal: 4096 pages in unavailable ranges Sep 13 01:20:31.027250 kernel: ACPI: PM-Timer IO Port: 0x1808 Sep 13 01:20:31.027256 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Sep 13 01:20:31.027261 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Sep 13 01:20:31.027267 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Sep 13 01:20:31.027273 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Sep 13 01:20:31.027279 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Sep 13 01:20:31.027284 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Sep 13 01:20:31.027290 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Sep 13 01:20:31.027295 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Sep 13 01:20:31.027301 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Sep 13 01:20:31.027306 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Sep 13 01:20:31.027311 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Sep 13 01:20:31.027317 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Sep 13 01:20:31.027323 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Sep 13 01:20:31.027329 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Sep 13 01:20:31.027334 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Sep 13 01:20:31.027342 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Sep 13 01:20:31.027347 kernel: IOAPIC[0]: apic_id 2, version 32, address 0xfec00000, GSI 0-119 Sep 13 01:20:31.027353 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 13 01:20:31.027358 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 13 01:20:31.027364 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 13 01:20:31.027369 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 13 01:20:31.027376 kernel: TSC deadline timer available Sep 13 01:20:31.027382 kernel: smpboot: Allowing 16 CPUs, 0 hotplug CPUs Sep 13 01:20:31.027387 kernel: [mem 0x90000000-0xdfffffff] available for PCI devices Sep 13 01:20:31.027393 kernel: Booting paravirtualized kernel on bare hardware Sep 13 01:20:31.027398 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 13 01:20:31.027404 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Sep 13 01:20:31.027409 kernel: percpu: Embedded 58 pages/cpu s197160 r8192 d32216 u262144 Sep 13 01:20:31.027415 kernel: pcpu-alloc: s197160 r8192 d32216 u262144 alloc=1*2097152 Sep 13 01:20:31.027420 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Sep 13 01:20:31.027427 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:20:31.027433 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 13 01:20:31.027438 kernel: random: crng init done Sep 13 01:20:31.027444 kernel: Dentry cache hash table entries: 4194304 (order: 13, 33554432 bytes, linear) Sep 13 01:20:31.027449 kernel: Inode-cache hash table entries: 2097152 (order: 12, 16777216 bytes, linear) Sep 13 01:20:31.027455 kernel: Fallback order for Node 0: 0 Sep 13 01:20:31.027460 kernel: Built 1 zonelists, mobility grouping on. Total pages: 8232416 Sep 13 01:20:31.027465 kernel: Policy zone: Normal Sep 13 01:20:31.027472 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 13 01:20:31.027477 kernel: software IO TLB: area num 16. Sep 13 01:20:31.027483 kernel: Memory: 32720304K/33452984K available (12288K kernel code, 2293K rwdata, 22744K rodata, 42884K init, 2312K bss, 732420K reserved, 0K cma-reserved) Sep 13 01:20:31.027489 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Sep 13 01:20:31.027494 kernel: ftrace: allocating 37974 entries in 149 pages Sep 13 01:20:31.027500 kernel: ftrace: allocated 149 pages with 4 groups Sep 13 01:20:31.027505 kernel: Dynamic Preempt: voluntary Sep 13 01:20:31.027511 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 13 01:20:31.027517 kernel: rcu: RCU event tracing is enabled. Sep 13 01:20:31.027523 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Sep 13 01:20:31.027529 kernel: Trampoline variant of Tasks RCU enabled. Sep 13 01:20:31.027534 kernel: Rude variant of Tasks RCU enabled. Sep 13 01:20:31.027540 kernel: Tracing variant of Tasks RCU enabled. Sep 13 01:20:31.027545 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 13 01:20:31.027551 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Sep 13 01:20:31.027556 kernel: NR_IRQS: 33024, nr_irqs: 2184, preallocated irqs: 16 Sep 13 01:20:31.027562 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 13 01:20:31.027567 kernel: Console: colour dummy device 80x25 Sep 13 01:20:31.027573 kernel: printk: console [tty0] enabled Sep 13 01:20:31.027579 kernel: printk: console [ttyS1] enabled Sep 13 01:20:31.027585 kernel: ACPI: Core revision 20230628 Sep 13 01:20:31.027590 kernel: hpet: HPET dysfunctional in PC10. Force disabled. Sep 13 01:20:31.027596 kernel: APIC: Switch to symmetric I/O mode setup Sep 13 01:20:31.027601 kernel: DMAR: Host address width 39 Sep 13 01:20:31.027606 kernel: DMAR: DRHD base: 0x000000fed91000 flags: 0x1 Sep 13 01:20:31.027612 kernel: DMAR: dmar0: reg_base_addr fed91000 ver 1:0 cap d2008c40660462 ecap f050da Sep 13 01:20:31.027617 kernel: DMAR: RMRR base: 0x0000008cf19000 end: 0x0000008d162fff Sep 13 01:20:31.027623 kernel: DMAR-IR: IOAPIC id 2 under DRHD base 0xfed91000 IOMMU 0 Sep 13 01:20:31.027629 kernel: DMAR-IR: HPET id 0 under DRHD base 0xfed91000 Sep 13 01:20:31.027635 kernel: DMAR-IR: Queued invalidation will be enabled to support x2apic and Intr-remapping. Sep 13 01:20:31.027640 kernel: DMAR-IR: Enabled IRQ remapping in x2apic mode Sep 13 01:20:31.027646 kernel: x2apic enabled Sep 13 01:20:31.027651 kernel: APIC: Switched APIC routing to: cluster x2apic Sep 13 01:20:31.027657 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x3101f59f5e6, max_idle_ns: 440795259996 ns Sep 13 01:20:31.027663 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 6799.81 BogoMIPS (lpj=3399906) Sep 13 01:20:31.027668 kernel: CPU0: Thermal monitoring enabled (TM1) Sep 13 01:20:31.027673 kernel: process: using mwait in idle threads Sep 13 01:20:31.027680 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Sep 13 01:20:31.027685 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Sep 13 01:20:31.027691 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 13 01:20:31.027696 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Sep 13 01:20:31.027701 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Sep 13 01:20:31.027707 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Sep 13 01:20:31.027712 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Sep 13 01:20:31.027718 kernel: RETBleed: Mitigation: Enhanced IBRS Sep 13 01:20:31.027723 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 13 01:20:31.027729 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 13 01:20:31.027734 kernel: TAA: Mitigation: TSX disabled Sep 13 01:20:31.027740 kernel: MMIO Stale Data: Mitigation: Clear CPU buffers Sep 13 01:20:31.027746 kernel: SRBDS: Mitigation: Microcode Sep 13 01:20:31.027751 kernel: GDS: Mitigation: Microcode Sep 13 01:20:31.027757 kernel: active return thunk: its_return_thunk Sep 13 01:20:31.027762 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 13 01:20:31.027767 kernel: VMSCAPE: Mitigation: IBPB before exit to userspace Sep 13 01:20:31.027773 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 13 01:20:31.027778 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 13 01:20:31.027784 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 13 01:20:31.027789 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Sep 13 01:20:31.027794 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Sep 13 01:20:31.027801 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 13 01:20:31.027806 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Sep 13 01:20:31.027812 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Sep 13 01:20:31.027817 kernel: x86/fpu: Enabled xstate features 0x1f, context size is 960 bytes, using 'compacted' format. Sep 13 01:20:31.027823 kernel: Freeing SMP alternatives memory: 32K Sep 13 01:20:31.027828 kernel: pid_max: default: 32768 minimum: 301 Sep 13 01:20:31.027834 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 13 01:20:31.027839 kernel: landlock: Up and running. Sep 13 01:20:31.027845 kernel: SELinux: Initializing. Sep 13 01:20:31.027850 kernel: Mount-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.027856 kernel: Mountpoint-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.027862 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Sep 13 01:20:31.027868 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:20:31.027873 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:20:31.027879 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:20:31.027884 kernel: Performance Events: PEBS fmt3+, Skylake events, 32-deep LBR, full-width counters, Intel PMU driver. Sep 13 01:20:31.027890 kernel: ... version: 4 Sep 13 01:20:31.027896 kernel: ... bit width: 48 Sep 13 01:20:31.027901 kernel: ... generic registers: 4 Sep 13 01:20:31.027906 kernel: ... value mask: 0000ffffffffffff Sep 13 01:20:31.027913 kernel: ... max period: 00007fffffffffff Sep 13 01:20:31.027919 kernel: ... fixed-purpose events: 3 Sep 13 01:20:31.027924 kernel: ... event mask: 000000070000000f Sep 13 01:20:31.027930 kernel: signal: max sigframe size: 2032 Sep 13 01:20:31.027935 kernel: Estimated ratio of average max frequency by base frequency (times 1024): 1445 Sep 13 01:20:31.027941 kernel: rcu: Hierarchical SRCU implementation. Sep 13 01:20:31.027946 kernel: rcu: Max phase no-delay instances is 400. Sep 13 01:20:31.027952 kernel: NMI watchdog: Enabled. Permanently consumes one hw-PMU counter. Sep 13 01:20:31.027957 kernel: smp: Bringing up secondary CPUs ... Sep 13 01:20:31.027963 kernel: smpboot: x86: Booting SMP configuration: Sep 13 01:20:31.027969 kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6 #7 #8 #9 #10 #11 #12 #13 #14 #15 Sep 13 01:20:31.027975 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 13 01:20:31.027981 kernel: smp: Brought up 1 node, 16 CPUs Sep 13 01:20:31.027986 kernel: smpboot: Max logical packages: 1 Sep 13 01:20:31.027992 kernel: smpboot: Total of 16 processors activated (108796.99 BogoMIPS) Sep 13 01:20:31.027997 kernel: devtmpfs: initialized Sep 13 01:20:31.028003 kernel: x86/mm: Memory block size: 128MB Sep 13 01:20:31.028008 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x825be000-0x825befff] (4096 bytes) Sep 13 01:20:31.028015 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x8c23c000-0x8c66dfff] (4399104 bytes) Sep 13 01:20:31.028020 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 13 01:20:31.028026 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Sep 13 01:20:31.028031 kernel: pinctrl core: initialized pinctrl subsystem Sep 13 01:20:31.028037 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 13 01:20:31.028042 kernel: audit: initializing netlink subsys (disabled) Sep 13 01:20:31.028048 kernel: audit: type=2000 audit(1757726425.039:1): state=initialized audit_enabled=0 res=1 Sep 13 01:20:31.028053 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 13 01:20:31.028059 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 13 01:20:31.028065 kernel: cpuidle: using governor menu Sep 13 01:20:31.028070 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 13 01:20:31.028076 kernel: dca service started, version 1.12.1 Sep 13 01:20:31.028081 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Sep 13 01:20:31.028087 kernel: PCI: Using configuration type 1 for base access Sep 13 01:20:31.028092 kernel: ENERGY_PERF_BIAS: Set to 'normal', was 'performance' Sep 13 01:20:31.028098 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 13 01:20:31.028103 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 13 01:20:31.028109 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 13 01:20:31.028115 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 13 01:20:31.028121 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 13 01:20:31.028126 kernel: ACPI: Added _OSI(Module Device) Sep 13 01:20:31.028132 kernel: ACPI: Added _OSI(Processor Device) Sep 13 01:20:31.028137 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 13 01:20:31.028143 kernel: ACPI: 12 ACPI AML tables successfully acquired and loaded Sep 13 01:20:31.028148 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028154 kernel: ACPI: SSDT 0xFFFF98DD41AF2400 000400 (v02 PmRef Cpu0Cst 00003001 INTL 20160527) Sep 13 01:20:31.028159 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028166 kernel: ACPI: SSDT 0xFFFF98DD41AEC800 000683 (v02 PmRef Cpu0Ist 00003000 INTL 20160527) Sep 13 01:20:31.028171 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028177 kernel: ACPI: SSDT 0xFFFF98DD40247100 0000F4 (v02 PmRef Cpu0Psd 00003000 INTL 20160527) Sep 13 01:20:31.028182 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028187 kernel: ACPI: SSDT 0xFFFF98DD41E5F800 0005FC (v02 PmRef ApIst 00003000 INTL 20160527) Sep 13 01:20:31.028193 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028198 kernel: ACPI: SSDT 0xFFFF98DD4012E000 000AB0 (v02 PmRef ApPsd 00003000 INTL 20160527) Sep 13 01:20:31.028204 kernel: ACPI: Dynamic OEM Table Load: Sep 13 01:20:31.028209 kernel: ACPI: SSDT 0xFFFF98DD41AF4C00 00030A (v02 PmRef ApCst 00003000 INTL 20160527) Sep 13 01:20:31.028215 kernel: ACPI: _OSC evaluated successfully for all CPUs Sep 13 01:20:31.028221 kernel: ACPI: Interpreter enabled Sep 13 01:20:31.028226 kernel: ACPI: PM: (supports S0 S5) Sep 13 01:20:31.028232 kernel: ACPI: Using IOAPIC for interrupt routing Sep 13 01:20:31.028237 kernel: HEST: Enabling Firmware First mode for corrected errors. Sep 13 01:20:31.028243 kernel: mce: [Firmware Bug]: Ignoring request to disable invalid MCA bank 14. Sep 13 01:20:31.028248 kernel: HEST: Table parsing has been initialized. Sep 13 01:20:31.028254 kernel: GHES: APEI firmware first mode is enabled by APEI bit and WHEA _OSC. Sep 13 01:20:31.028259 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 13 01:20:31.028265 kernel: PCI: Ignoring E820 reservations for host bridge windows Sep 13 01:20:31.028271 kernel: ACPI: Enabled 9 GPEs in block 00 to 7F Sep 13 01:20:31.028277 kernel: ACPI: \_SB_.PCI0.XDCI.USBC: New power resource Sep 13 01:20:31.028282 kernel: ACPI: \_SB_.PCI0.SAT0.VOL0.V0PR: New power resource Sep 13 01:20:31.028288 kernel: ACPI: \_SB_.PCI0.SAT0.VOL1.V1PR: New power resource Sep 13 01:20:31.028293 kernel: ACPI: \_SB_.PCI0.SAT0.VOL2.V2PR: New power resource Sep 13 01:20:31.028299 kernel: ACPI: \_SB_.PCI0.CNVW.WRST: New power resource Sep 13 01:20:31.028304 kernel: ACPI: \_TZ_.FN00: New power resource Sep 13 01:20:31.028310 kernel: ACPI: \_TZ_.FN01: New power resource Sep 13 01:20:31.028315 kernel: ACPI: \_TZ_.FN02: New power resource Sep 13 01:20:31.028322 kernel: ACPI: \_TZ_.FN03: New power resource Sep 13 01:20:31.028327 kernel: ACPI: \_TZ_.FN04: New power resource Sep 13 01:20:31.028333 kernel: ACPI: \PIN_: New power resource Sep 13 01:20:31.028340 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-fe]) Sep 13 01:20:31.028415 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 13 01:20:31.028470 kernel: acpi PNP0A08:00: _OSC: platform does not support [AER] Sep 13 01:20:31.028520 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability LTR] Sep 13 01:20:31.028530 kernel: PCI host bridge to bus 0000:00 Sep 13 01:20:31.028580 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 13 01:20:31.028624 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 13 01:20:31.028668 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 13 01:20:31.028711 kernel: pci_bus 0000:00: root bus resource [mem 0x90000000-0xdfffffff window] Sep 13 01:20:31.028753 kernel: pci_bus 0000:00: root bus resource [mem 0xfc800000-0xfe7fffff window] Sep 13 01:20:31.028796 kernel: pci_bus 0000:00: root bus resource [bus 00-fe] Sep 13 01:20:31.028859 kernel: pci 0000:00:00.0: [8086:3e31] type 00 class 0x060000 Sep 13 01:20:31.028916 kernel: pci 0000:00:01.0: [8086:1901] type 01 class 0x060400 Sep 13 01:20:31.028966 kernel: pci 0000:00:01.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.029019 kernel: pci 0000:00:08.0: [8086:1911] type 00 class 0x088000 Sep 13 01:20:31.029069 kernel: pci 0000:00:08.0: reg 0x10: [mem 0x9551f000-0x9551ffff 64bit] Sep 13 01:20:31.029120 kernel: pci 0000:00:12.0: [8086:a379] type 00 class 0x118000 Sep 13 01:20:31.029172 kernel: pci 0000:00:12.0: reg 0x10: [mem 0x9551e000-0x9551efff 64bit] Sep 13 01:20:31.029225 kernel: pci 0000:00:14.0: [8086:a36d] type 00 class 0x0c0330 Sep 13 01:20:31.029274 kernel: pci 0000:00:14.0: reg 0x10: [mem 0x95500000-0x9550ffff 64bit] Sep 13 01:20:31.029322 kernel: pci 0000:00:14.0: PME# supported from D3hot D3cold Sep 13 01:20:31.029380 kernel: pci 0000:00:14.2: [8086:a36f] type 00 class 0x050000 Sep 13 01:20:31.029429 kernel: pci 0000:00:14.2: reg 0x10: [mem 0x95512000-0x95513fff 64bit] Sep 13 01:20:31.029481 kernel: pci 0000:00:14.2: reg 0x18: [mem 0x9551d000-0x9551dfff 64bit] Sep 13 01:20:31.029533 kernel: pci 0000:00:15.0: [8086:a368] type 00 class 0x0c8000 Sep 13 01:20:31.029582 kernel: pci 0000:00:15.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Sep 13 01:20:31.029638 kernel: pci 0000:00:15.1: [8086:a369] type 00 class 0x0c8000 Sep 13 01:20:31.029688 kernel: pci 0000:00:15.1: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Sep 13 01:20:31.029741 kernel: pci 0000:00:16.0: [8086:a360] type 00 class 0x078000 Sep 13 01:20:31.029792 kernel: pci 0000:00:16.0: reg 0x10: [mem 0x9551a000-0x9551afff 64bit] Sep 13 01:20:31.029843 kernel: pci 0000:00:16.0: PME# supported from D3hot Sep 13 01:20:31.029903 kernel: pci 0000:00:16.1: [8086:a361] type 00 class 0x078000 Sep 13 01:20:31.029955 kernel: pci 0000:00:16.1: reg 0x10: [mem 0x95519000-0x95519fff 64bit] Sep 13 01:20:31.030003 kernel: pci 0000:00:16.1: PME# supported from D3hot Sep 13 01:20:31.030055 kernel: pci 0000:00:16.4: [8086:a364] type 00 class 0x078000 Sep 13 01:20:31.030104 kernel: pci 0000:00:16.4: reg 0x10: [mem 0x95518000-0x95518fff 64bit] Sep 13 01:20:31.030155 kernel: pci 0000:00:16.4: PME# supported from D3hot Sep 13 01:20:31.030208 kernel: pci 0000:00:17.0: [8086:a352] type 00 class 0x010601 Sep 13 01:20:31.030258 kernel: pci 0000:00:17.0: reg 0x10: [mem 0x95510000-0x95511fff] Sep 13 01:20:31.030306 kernel: pci 0000:00:17.0: reg 0x14: [mem 0x95517000-0x955170ff] Sep 13 01:20:31.030357 kernel: pci 0000:00:17.0: reg 0x18: [io 0x6050-0x6057] Sep 13 01:20:31.030405 kernel: pci 0000:00:17.0: reg 0x1c: [io 0x6040-0x6043] Sep 13 01:20:31.030454 kernel: pci 0000:00:17.0: reg 0x20: [io 0x6020-0x603f] Sep 13 01:20:31.030506 kernel: pci 0000:00:17.0: reg 0x24: [mem 0x95516000-0x955167ff] Sep 13 01:20:31.030554 kernel: pci 0000:00:17.0: PME# supported from D3hot Sep 13 01:20:31.030610 kernel: pci 0000:00:1b.0: [8086:a340] type 01 class 0x060400 Sep 13 01:20:31.030659 kernel: pci 0000:00:1b.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.030718 kernel: pci 0000:00:1b.4: [8086:a32c] type 01 class 0x060400 Sep 13 01:20:31.030768 kernel: pci 0000:00:1b.4: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.030821 kernel: pci 0000:00:1b.5: [8086:a32d] type 01 class 0x060400 Sep 13 01:20:31.030870 kernel: pci 0000:00:1b.5: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.030924 kernel: pci 0000:00:1c.0: [8086:a338] type 01 class 0x060400 Sep 13 01:20:31.030975 kernel: pci 0000:00:1c.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.031033 kernel: pci 0000:00:1c.3: [8086:a33b] type 01 class 0x060400 Sep 13 01:20:31.031082 kernel: pci 0000:00:1c.3: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.031136 kernel: pci 0000:00:1e.0: [8086:a328] type 00 class 0x078000 Sep 13 01:20:31.031184 kernel: pci 0000:00:1e.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Sep 13 01:20:31.031238 kernel: pci 0000:00:1f.0: [8086:a309] type 00 class 0x060100 Sep 13 01:20:31.031290 kernel: pci 0000:00:1f.4: [8086:a323] type 00 class 0x0c0500 Sep 13 01:20:31.031346 kernel: pci 0000:00:1f.4: reg 0x10: [mem 0x95514000-0x955140ff 64bit] Sep 13 01:20:31.031396 kernel: pci 0000:00:1f.4: reg 0x20: [io 0xefa0-0xefbf] Sep 13 01:20:31.031451 kernel: pci 0000:00:1f.5: [8086:a324] type 00 class 0x0c8000 Sep 13 01:20:31.031500 kernel: pci 0000:00:1f.5: reg 0x10: [mem 0xfe010000-0xfe010fff] Sep 13 01:20:31.031555 kernel: pci 0000:01:00.0: [15b3:1015] type 00 class 0x020000 Sep 13 01:20:31.031606 kernel: pci 0000:01:00.0: reg 0x10: [mem 0x92000000-0x93ffffff 64bit pref] Sep 13 01:20:31.031659 kernel: pci 0000:01:00.0: reg 0x30: [mem 0x95200000-0x952fffff pref] Sep 13 01:20:31.031709 kernel: pci 0000:01:00.0: PME# supported from D3cold Sep 13 01:20:31.031759 kernel: pci 0000:01:00.0: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Sep 13 01:20:31.031810 kernel: pci 0000:01:00.0: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Sep 13 01:20:31.031865 kernel: pci 0000:01:00.1: [15b3:1015] type 00 class 0x020000 Sep 13 01:20:31.031915 kernel: pci 0000:01:00.1: reg 0x10: [mem 0x90000000-0x91ffffff 64bit pref] Sep 13 01:20:31.031966 kernel: pci 0000:01:00.1: reg 0x30: [mem 0x95100000-0x951fffff pref] Sep 13 01:20:31.032017 kernel: pci 0000:01:00.1: PME# supported from D3cold Sep 13 01:20:31.032068 kernel: pci 0000:01:00.1: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Sep 13 01:20:31.032117 kernel: pci 0000:01:00.1: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Sep 13 01:20:31.032168 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Sep 13 01:20:31.032217 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Sep 13 01:20:31.032267 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Sep 13 01:20:31.032315 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Sep 13 01:20:31.032375 kernel: pci 0000:03:00.0: working around ROM BAR overlap defect Sep 13 01:20:31.032429 kernel: pci 0000:03:00.0: [8086:1533] type 00 class 0x020000 Sep 13 01:20:31.032480 kernel: pci 0000:03:00.0: reg 0x10: [mem 0x95400000-0x9547ffff] Sep 13 01:20:31.032530 kernel: pci 0000:03:00.0: reg 0x18: [io 0x5000-0x501f] Sep 13 01:20:31.032579 kernel: pci 0000:03:00.0: reg 0x1c: [mem 0x95480000-0x95483fff] Sep 13 01:20:31.032631 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.032681 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Sep 13 01:20:31.032731 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Sep 13 01:20:31.032783 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Sep 13 01:20:31.032839 kernel: pci 0000:04:00.0: working around ROM BAR overlap defect Sep 13 01:20:31.032891 kernel: pci 0000:04:00.0: [8086:1533] type 00 class 0x020000 Sep 13 01:20:31.032941 kernel: pci 0000:04:00.0: reg 0x10: [mem 0x95300000-0x9537ffff] Sep 13 01:20:31.032993 kernel: pci 0000:04:00.0: reg 0x18: [io 0x4000-0x401f] Sep 13 01:20:31.033043 kernel: pci 0000:04:00.0: reg 0x1c: [mem 0x95380000-0x95383fff] Sep 13 01:20:31.033094 kernel: pci 0000:04:00.0: PME# supported from D0 D3hot D3cold Sep 13 01:20:31.033146 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Sep 13 01:20:31.033196 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Sep 13 01:20:31.033245 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Sep 13 01:20:31.033295 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Sep 13 01:20:31.033355 kernel: pci 0000:06:00.0: [1a03:1150] type 01 class 0x060400 Sep 13 01:20:31.033406 kernel: pci 0000:06:00.0: enabling Extended Tags Sep 13 01:20:31.033457 kernel: pci 0000:06:00.0: supports D1 D2 Sep 13 01:20:31.033506 kernel: pci 0000:06:00.0: PME# supported from D0 D1 D2 D3hot D3cold Sep 13 01:20:31.033559 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Sep 13 01:20:31.033608 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.033658 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.033713 kernel: pci_bus 0000:07: extended config space not accessible Sep 13 01:20:31.033771 kernel: pci 0000:07:00.0: [1a03:2000] type 00 class 0x030000 Sep 13 01:20:31.033824 kernel: pci 0000:07:00.0: reg 0x10: [mem 0x94000000-0x94ffffff] Sep 13 01:20:31.033875 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x95000000-0x9501ffff] Sep 13 01:20:31.033931 kernel: pci 0000:07:00.0: reg 0x18: [io 0x3000-0x307f] Sep 13 01:20:31.033982 kernel: pci 0000:07:00.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 13 01:20:31.034035 kernel: pci 0000:07:00.0: supports D1 D2 Sep 13 01:20:31.034088 kernel: pci 0000:07:00.0: PME# supported from D0 D1 D2 D3hot D3cold Sep 13 01:20:31.034138 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Sep 13 01:20:31.034188 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.034238 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.034246 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 0 Sep 13 01:20:31.034254 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 1 Sep 13 01:20:31.034260 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 0 Sep 13 01:20:31.034267 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 0 Sep 13 01:20:31.034273 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 0 Sep 13 01:20:31.034279 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 0 Sep 13 01:20:31.034285 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 0 Sep 13 01:20:31.034290 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 0 Sep 13 01:20:31.034296 kernel: iommu: Default domain type: Translated Sep 13 01:20:31.034302 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 13 01:20:31.034309 kernel: PCI: Using ACPI for IRQ routing Sep 13 01:20:31.034315 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 13 01:20:31.034320 kernel: e820: reserve RAM buffer [mem 0x00099800-0x0009ffff] Sep 13 01:20:31.034326 kernel: e820: reserve RAM buffer [mem 0x825be000-0x83ffffff] Sep 13 01:20:31.034332 kernel: e820: reserve RAM buffer [mem 0x8afce000-0x8bffffff] Sep 13 01:20:31.034339 kernel: e820: reserve RAM buffer [mem 0x8c23c000-0x8fffffff] Sep 13 01:20:31.034345 kernel: e820: reserve RAM buffer [mem 0x8ef00000-0x8fffffff] Sep 13 01:20:31.034351 kernel: e820: reserve RAM buffer [mem 0x86f000000-0x86fffffff] Sep 13 01:20:31.034403 kernel: pci 0000:07:00.0: vgaarb: setting as boot VGA device Sep 13 01:20:31.034457 kernel: pci 0000:07:00.0: vgaarb: bridge control possible Sep 13 01:20:31.034510 kernel: pci 0000:07:00.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 13 01:20:31.034518 kernel: vgaarb: loaded Sep 13 01:20:31.034524 kernel: clocksource: Switched to clocksource tsc-early Sep 13 01:20:31.034530 kernel: VFS: Disk quotas dquot_6.6.0 Sep 13 01:20:31.034536 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 13 01:20:31.034542 kernel: pnp: PnP ACPI init Sep 13 01:20:31.034592 kernel: system 00:00: [mem 0x40000000-0x403fffff] has been reserved Sep 13 01:20:31.034643 kernel: pnp 00:02: [dma 0 disabled] Sep 13 01:20:31.034693 kernel: pnp 00:03: [dma 0 disabled] Sep 13 01:20:31.034745 kernel: system 00:04: [io 0x0680-0x069f] has been reserved Sep 13 01:20:31.034791 kernel: system 00:04: [io 0x164e-0x164f] has been reserved Sep 13 01:20:31.034839 kernel: system 00:05: [mem 0xfed10000-0xfed17fff] has been reserved Sep 13 01:20:31.034884 kernel: system 00:05: [mem 0xfed18000-0xfed18fff] has been reserved Sep 13 01:20:31.034931 kernel: system 00:05: [mem 0xfed19000-0xfed19fff] has been reserved Sep 13 01:20:31.034976 kernel: system 00:05: [mem 0xe0000000-0xefffffff] has been reserved Sep 13 01:20:31.035020 kernel: system 00:05: [mem 0xfed20000-0xfed3ffff] has been reserved Sep 13 01:20:31.035068 kernel: system 00:05: [mem 0xfed90000-0xfed93fff] could not be reserved Sep 13 01:20:31.035113 kernel: system 00:05: [mem 0xfed45000-0xfed8ffff] has been reserved Sep 13 01:20:31.035158 kernel: system 00:05: [mem 0xfee00000-0xfeefffff] could not be reserved Sep 13 01:20:31.035206 kernel: system 00:06: [io 0x1800-0x18fe] could not be reserved Sep 13 01:20:31.035255 kernel: system 00:06: [mem 0xfd000000-0xfd69ffff] has been reserved Sep 13 01:20:31.035299 kernel: system 00:06: [mem 0xfd6c0000-0xfd6cffff] has been reserved Sep 13 01:20:31.035346 kernel: system 00:06: [mem 0xfd6f0000-0xfdffffff] has been reserved Sep 13 01:20:31.035391 kernel: system 00:06: [mem 0xfe000000-0xfe01ffff] could not be reserved Sep 13 01:20:31.035435 kernel: system 00:06: [mem 0xfe200000-0xfe7fffff] has been reserved Sep 13 01:20:31.035480 kernel: system 00:06: [mem 0xff000000-0xffffffff] has been reserved Sep 13 01:20:31.035528 kernel: system 00:07: [io 0x2000-0x20fe] has been reserved Sep 13 01:20:31.035538 kernel: pnp: PnP ACPI: found 9 devices Sep 13 01:20:31.035544 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 13 01:20:31.035550 kernel: NET: Registered PF_INET protocol family Sep 13 01:20:31.035556 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 01:20:31.035562 kernel: tcp_listen_portaddr_hash hash table entries: 16384 (order: 6, 262144 bytes, linear) Sep 13 01:20:31.035568 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 13 01:20:31.035574 kernel: TCP established hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 01:20:31.035579 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Sep 13 01:20:31.035586 kernel: TCP: Hash tables configured (established 262144 bind 65536) Sep 13 01:20:31.035592 kernel: UDP hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.035598 kernel: UDP-Lite hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 13 01:20:31.035604 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 13 01:20:31.035610 kernel: NET: Registered PF_XDP protocol family Sep 13 01:20:31.035659 kernel: pci 0000:00:15.0: BAR 0: assigned [mem 0x95515000-0x95515fff 64bit] Sep 13 01:20:31.035709 kernel: pci 0000:00:15.1: BAR 0: assigned [mem 0x9551b000-0x9551bfff 64bit] Sep 13 01:20:31.035760 kernel: pci 0000:00:1e.0: BAR 0: assigned [mem 0x9551c000-0x9551cfff 64bit] Sep 13 01:20:31.035811 kernel: pci 0000:01:00.0: BAR 7: no space for [mem size 0x00800000 64bit pref] Sep 13 01:20:31.035866 kernel: pci 0000:01:00.0: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Sep 13 01:20:31.035917 kernel: pci 0000:01:00.1: BAR 7: no space for [mem size 0x00800000 64bit pref] Sep 13 01:20:31.035968 kernel: pci 0000:01:00.1: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Sep 13 01:20:31.036018 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Sep 13 01:20:31.036067 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Sep 13 01:20:31.036116 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Sep 13 01:20:31.036168 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Sep 13 01:20:31.036220 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Sep 13 01:20:31.036268 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Sep 13 01:20:31.036318 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Sep 13 01:20:31.036370 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Sep 13 01:20:31.036420 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Sep 13 01:20:31.036470 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Sep 13 01:20:31.036520 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Sep 13 01:20:31.036569 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Sep 13 01:20:31.036619 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.036670 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.036719 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Sep 13 01:20:31.036769 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Sep 13 01:20:31.036819 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Sep 13 01:20:31.036865 kernel: pci_bus 0000:00: Some PCI device resources are unassigned, try booting with pci=realloc Sep 13 01:20:31.036911 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 13 01:20:31.036956 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 13 01:20:31.036998 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 13 01:20:31.037042 kernel: pci_bus 0000:00: resource 7 [mem 0x90000000-0xdfffffff window] Sep 13 01:20:31.037084 kernel: pci_bus 0000:00: resource 8 [mem 0xfc800000-0xfe7fffff window] Sep 13 01:20:31.037134 kernel: pci_bus 0000:01: resource 1 [mem 0x95100000-0x952fffff] Sep 13 01:20:31.037180 kernel: pci_bus 0000:01: resource 2 [mem 0x90000000-0x93ffffff 64bit pref] Sep 13 01:20:31.037232 kernel: pci_bus 0000:03: resource 0 [io 0x5000-0x5fff] Sep 13 01:20:31.037277 kernel: pci_bus 0000:03: resource 1 [mem 0x95400000-0x954fffff] Sep 13 01:20:31.037330 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Sep 13 01:20:31.037377 kernel: pci_bus 0000:04: resource 1 [mem 0x95300000-0x953fffff] Sep 13 01:20:31.037427 kernel: pci_bus 0000:06: resource 0 [io 0x3000-0x3fff] Sep 13 01:20:31.037473 kernel: pci_bus 0000:06: resource 1 [mem 0x94000000-0x950fffff] Sep 13 01:20:31.037520 kernel: pci_bus 0000:07: resource 0 [io 0x3000-0x3fff] Sep 13 01:20:31.037569 kernel: pci_bus 0000:07: resource 1 [mem 0x94000000-0x950fffff] Sep 13 01:20:31.037577 kernel: PCI: CLS 64 bytes, default 64 Sep 13 01:20:31.037583 kernel: DMAR: No ATSR found Sep 13 01:20:31.037589 kernel: DMAR: No SATC found Sep 13 01:20:31.037595 kernel: DMAR: dmar0: Using Queued invalidation Sep 13 01:20:31.037644 kernel: pci 0000:00:00.0: Adding to iommu group 0 Sep 13 01:20:31.037694 kernel: pci 0000:00:01.0: Adding to iommu group 1 Sep 13 01:20:31.037744 kernel: pci 0000:00:08.0: Adding to iommu group 2 Sep 13 01:20:31.037796 kernel: pci 0000:00:12.0: Adding to iommu group 3 Sep 13 01:20:31.037845 kernel: pci 0000:00:14.0: Adding to iommu group 4 Sep 13 01:20:31.037896 kernel: pci 0000:00:14.2: Adding to iommu group 4 Sep 13 01:20:31.037944 kernel: pci 0000:00:15.0: Adding to iommu group 5 Sep 13 01:20:31.037993 kernel: pci 0000:00:15.1: Adding to iommu group 5 Sep 13 01:20:31.038042 kernel: pci 0000:00:16.0: Adding to iommu group 6 Sep 13 01:20:31.038091 kernel: pci 0000:00:16.1: Adding to iommu group 6 Sep 13 01:20:31.038140 kernel: pci 0000:00:16.4: Adding to iommu group 6 Sep 13 01:20:31.038191 kernel: pci 0000:00:17.0: Adding to iommu group 7 Sep 13 01:20:31.038239 kernel: pci 0000:00:1b.0: Adding to iommu group 8 Sep 13 01:20:31.038288 kernel: pci 0000:00:1b.4: Adding to iommu group 9 Sep 13 01:20:31.038336 kernel: pci 0000:00:1b.5: Adding to iommu group 10 Sep 13 01:20:31.038390 kernel: pci 0000:00:1c.0: Adding to iommu group 11 Sep 13 01:20:31.038440 kernel: pci 0000:00:1c.3: Adding to iommu group 12 Sep 13 01:20:31.038488 kernel: pci 0000:00:1e.0: Adding to iommu group 13 Sep 13 01:20:31.038538 kernel: pci 0000:00:1f.0: Adding to iommu group 14 Sep 13 01:20:31.038589 kernel: pci 0000:00:1f.4: Adding to iommu group 14 Sep 13 01:20:31.038639 kernel: pci 0000:00:1f.5: Adding to iommu group 14 Sep 13 01:20:31.038689 kernel: pci 0000:01:00.0: Adding to iommu group 1 Sep 13 01:20:31.038740 kernel: pci 0000:01:00.1: Adding to iommu group 1 Sep 13 01:20:31.038791 kernel: pci 0000:03:00.0: Adding to iommu group 15 Sep 13 01:20:31.038841 kernel: pci 0000:04:00.0: Adding to iommu group 16 Sep 13 01:20:31.038892 kernel: pci 0000:06:00.0: Adding to iommu group 17 Sep 13 01:20:31.038946 kernel: pci 0000:07:00.0: Adding to iommu group 17 Sep 13 01:20:31.038956 kernel: DMAR: Intel(R) Virtualization Technology for Directed I/O Sep 13 01:20:31.038962 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Sep 13 01:20:31.038968 kernel: software IO TLB: mapped [mem 0x0000000086fce000-0x000000008afce000] (64MB) Sep 13 01:20:31.038974 kernel: RAPL PMU: API unit is 2^-32 Joules, 3 fixed counters, 655360 ms ovfl timer Sep 13 01:20:31.038980 kernel: RAPL PMU: hw unit of domain pp0-core 2^-14 Joules Sep 13 01:20:31.038986 kernel: RAPL PMU: hw unit of domain package 2^-14 Joules Sep 13 01:20:31.038992 kernel: RAPL PMU: hw unit of domain dram 2^-14 Joules Sep 13 01:20:31.039044 kernel: platform rtc_cmos: registered platform RTC device (no PNP device found) Sep 13 01:20:31.039053 kernel: Initialise system trusted keyrings Sep 13 01:20:31.039061 kernel: workingset: timestamp_bits=39 max_order=23 bucket_order=0 Sep 13 01:20:31.039066 kernel: Key type asymmetric registered Sep 13 01:20:31.039072 kernel: Asymmetric key parser 'x509' registered Sep 13 01:20:31.039078 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 13 01:20:31.039084 kernel: io scheduler mq-deadline registered Sep 13 01:20:31.039089 kernel: io scheduler kyber registered Sep 13 01:20:31.039095 kernel: io scheduler bfq registered Sep 13 01:20:31.039143 kernel: pcieport 0000:00:01.0: PME: Signaling with IRQ 121 Sep 13 01:20:31.039193 kernel: pcieport 0000:00:1b.0: PME: Signaling with IRQ 122 Sep 13 01:20:31.039245 kernel: pcieport 0000:00:1b.4: PME: Signaling with IRQ 123 Sep 13 01:20:31.039294 kernel: pcieport 0000:00:1b.5: PME: Signaling with IRQ 124 Sep 13 01:20:31.039346 kernel: pcieport 0000:00:1c.0: PME: Signaling with IRQ 125 Sep 13 01:20:31.039397 kernel: pcieport 0000:00:1c.3: PME: Signaling with IRQ 126 Sep 13 01:20:31.039450 kernel: thermal LNXTHERM:00: registered as thermal_zone0 Sep 13 01:20:31.039459 kernel: ACPI: thermal: Thermal Zone [TZ00] (28 C) Sep 13 01:20:31.039465 kernel: ERST: Error Record Serialization Table (ERST) support is initialized. Sep 13 01:20:31.039473 kernel: pstore: Using crash dump compression: deflate Sep 13 01:20:31.039479 kernel: pstore: Registered erst as persistent store backend Sep 13 01:20:31.039485 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 13 01:20:31.039491 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 13 01:20:31.039497 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 13 01:20:31.039503 kernel: 00:03: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Sep 13 01:20:31.039509 kernel: hpet_acpi_add: no address or irqs in _CRS Sep 13 01:20:31.039560 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x1B, rev-id 16) Sep 13 01:20:31.039570 kernel: i8042: PNP: No PS/2 controller found. Sep 13 01:20:31.039615 kernel: rtc_cmos rtc_cmos: RTC can wake from S4 Sep 13 01:20:31.039661 kernel: rtc_cmos rtc_cmos: registered as rtc0 Sep 13 01:20:31.039706 kernel: rtc_cmos rtc_cmos: setting system clock to 2025-09-13T01:20:29 UTC (1757726429) Sep 13 01:20:31.039752 kernel: rtc_cmos rtc_cmos: alarms up to one month, y3k, 114 bytes nvram Sep 13 01:20:31.039760 kernel: intel_pstate: Intel P-state driver initializing Sep 13 01:20:31.039767 kernel: intel_pstate: Disabling energy efficiency optimization Sep 13 01:20:31.039772 kernel: intel_pstate: HWP enabled Sep 13 01:20:31.039780 kernel: vesafb: mode is 1024x768x8, linelength=1024, pages=0 Sep 13 01:20:31.039786 kernel: vesafb: scrolling: redraw Sep 13 01:20:31.039792 kernel: vesafb: Pseudocolor: size=0:8:8:8, shift=0:0:0:0 Sep 13 01:20:31.039798 kernel: vesafb: framebuffer at 0x94000000, mapped to 0x0000000067ea91ba, using 768k, total 768k Sep 13 01:20:31.039803 kernel: Console: switching to colour frame buffer device 128x48 Sep 13 01:20:31.039809 kernel: fb0: VESA VGA frame buffer device Sep 13 01:20:31.039815 kernel: NET: Registered PF_INET6 protocol family Sep 13 01:20:31.039821 kernel: Segment Routing with IPv6 Sep 13 01:20:31.039827 kernel: In-situ OAM (IOAM) with IPv6 Sep 13 01:20:31.039833 kernel: NET: Registered PF_PACKET protocol family Sep 13 01:20:31.039839 kernel: Key type dns_resolver registered Sep 13 01:20:31.039845 kernel: microcode: Current revision: 0x00000102 Sep 13 01:20:31.039851 kernel: microcode: Microcode Update Driver: v2.2. Sep 13 01:20:31.039856 kernel: IPI shorthand broadcast: enabled Sep 13 01:20:31.039862 kernel: sched_clock: Marking stable (1904000715, 1379361895)->(4413394452, -1130031842) Sep 13 01:20:31.039868 kernel: registered taskstats version 1 Sep 13 01:20:31.039874 kernel: Loading compiled-in X.509 certificates Sep 13 01:20:31.039880 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.106-flatcar: 1274e0c573ac8d09163d6bc6d1ee1445fb2f8cc6' Sep 13 01:20:31.039885 kernel: Key type .fscrypt registered Sep 13 01:20:31.039892 kernel: Key type fscrypt-provisioning registered Sep 13 01:20:31.039898 kernel: ima: Allocated hash algorithm: sha1 Sep 13 01:20:31.039904 kernel: ima: No architecture policies found Sep 13 01:20:31.039909 kernel: clk: Disabling unused clocks Sep 13 01:20:31.039915 kernel: Freeing unused kernel image (initmem) memory: 42884K Sep 13 01:20:31.039921 kernel: Write protecting the kernel read-only data: 36864k Sep 13 01:20:31.039927 kernel: Freeing unused kernel image (rodata/data gap) memory: 1832K Sep 13 01:20:31.039933 kernel: Run /init as init process Sep 13 01:20:31.039939 kernel: with arguments: Sep 13 01:20:31.039945 kernel: /init Sep 13 01:20:31.039951 kernel: with environment: Sep 13 01:20:31.039957 kernel: HOME=/ Sep 13 01:20:31.039962 kernel: TERM=linux Sep 13 01:20:31.039968 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 13 01:20:31.039975 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 13 01:20:31.039982 systemd[1]: Detected architecture x86-64. Sep 13 01:20:31.039990 systemd[1]: Running in initrd. Sep 13 01:20:31.039996 systemd[1]: No hostname configured, using default hostname. Sep 13 01:20:31.040002 systemd[1]: Hostname set to . Sep 13 01:20:31.040007 systemd[1]: Initializing machine ID from random generator. Sep 13 01:20:31.040014 systemd[1]: Queued start job for default target initrd.target. Sep 13 01:20:31.040020 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:20:31.040026 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:20:31.040032 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 13 01:20:31.040040 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 13 01:20:31.040046 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 13 01:20:31.040052 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 13 01:20:31.040059 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 13 01:20:31.040065 kernel: tsc: Refined TSC clocksource calibration: 3407.999 MHz Sep 13 01:20:31.040071 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd336761, max_idle_ns: 440795243819 ns Sep 13 01:20:31.040077 kernel: clocksource: Switched to clocksource tsc Sep 13 01:20:31.040084 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 13 01:20:31.040090 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:20:31.040096 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:20:31.040102 systemd[1]: Reached target paths.target - Path Units. Sep 13 01:20:31.040109 systemd[1]: Reached target slices.target - Slice Units. Sep 13 01:20:31.040115 systemd[1]: Reached target swap.target - Swaps. Sep 13 01:20:31.040121 systemd[1]: Reached target timers.target - Timer Units. Sep 13 01:20:31.040127 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 13 01:20:31.040134 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 13 01:20:31.040140 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 13 01:20:31.040146 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 13 01:20:31.040152 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:20:31.040158 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 13 01:20:31.040164 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:20:31.040171 systemd[1]: Reached target sockets.target - Socket Units. Sep 13 01:20:31.040177 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 13 01:20:31.040183 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 13 01:20:31.040190 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 13 01:20:31.040196 systemd[1]: Starting systemd-fsck-usr.service... Sep 13 01:20:31.040202 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 13 01:20:31.040219 systemd-journald[265]: Collecting audit messages is disabled. Sep 13 01:20:31.040234 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 13 01:20:31.040241 systemd-journald[265]: Journal started Sep 13 01:20:31.040254 systemd-journald[265]: Runtime Journal (/run/log/journal/376e6e56161f4d3e8ac7e283b728153a) is 8.0M, max 639.9M, 631.9M free. Sep 13 01:20:31.054384 systemd-modules-load[266]: Inserted module 'overlay' Sep 13 01:20:31.076343 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:20:31.104356 systemd[1]: Started systemd-journald.service - Journal Service. Sep 13 01:20:31.104844 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 13 01:20:31.166910 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 13 01:20:31.166926 kernel: Bridge firewalling registered Sep 13 01:20:31.161526 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:20:31.166901 systemd-modules-load[266]: Inserted module 'br_netfilter' Sep 13 01:20:31.186703 systemd[1]: Finished systemd-fsck-usr.service. Sep 13 01:20:31.197746 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 13 01:20:31.226660 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:31.256704 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:20:31.269260 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 13 01:20:31.270872 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 13 01:20:31.272515 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 13 01:20:31.277093 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 13 01:20:31.277793 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 13 01:20:31.277891 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:20:31.278664 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:20:31.279536 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 13 01:20:31.292628 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:31.296957 systemd-resolved[302]: Positive Trust Anchors: Sep 13 01:20:31.296963 systemd-resolved[302]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 01:20:31.296985 systemd-resolved[302]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 13 01:20:31.298591 systemd-resolved[302]: Defaulting to hostname 'linux'. Sep 13 01:20:31.302818 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 13 01:20:31.334687 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:20:31.352718 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:20:31.388720 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 13 01:20:31.502335 dracut-cmdline[307]: dracut-dracut-053 Sep 13 01:20:31.509565 dracut-cmdline[307]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:20:31.703375 kernel: SCSI subsystem initialized Sep 13 01:20:31.726371 kernel: Loading iSCSI transport class v2.0-870. Sep 13 01:20:31.749370 kernel: iscsi: registered transport (tcp) Sep 13 01:20:31.782132 kernel: iscsi: registered transport (qla4xxx) Sep 13 01:20:31.782150 kernel: QLogic iSCSI HBA Driver Sep 13 01:20:31.815193 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 13 01:20:31.835664 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 13 01:20:31.923048 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 13 01:20:31.923068 kernel: device-mapper: uevent: version 1.0.3 Sep 13 01:20:31.942633 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 13 01:20:32.001448 kernel: raid6: avx2x4 gen() 53433 MB/s Sep 13 01:20:32.033375 kernel: raid6: avx2x2 gen() 53370 MB/s Sep 13 01:20:32.069757 kernel: raid6: avx2x1 gen() 45332 MB/s Sep 13 01:20:32.069773 kernel: raid6: using algorithm avx2x4 gen() 53433 MB/s Sep 13 01:20:32.116786 kernel: raid6: .... xor() 17429 MB/s, rmw enabled Sep 13 01:20:32.116806 kernel: raid6: using avx2x2 recovery algorithm Sep 13 01:20:32.157387 kernel: xor: automatically using best checksumming function avx Sep 13 01:20:32.271374 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 13 01:20:32.276849 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 13 01:20:32.310691 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:20:32.317711 systemd-udevd[495]: Using default interface naming scheme 'v255'. Sep 13 01:20:32.321556 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:20:32.355534 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 13 01:20:32.421934 dracut-pre-trigger[506]: rd.md=0: removing MD RAID activation Sep 13 01:20:32.450014 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 13 01:20:32.468580 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 13 01:20:32.542318 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:20:32.587102 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 13 01:20:32.587138 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 13 01:20:32.553316 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 13 01:20:32.590501 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 01:20:32.590532 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:32.603389 kernel: cryptd: max_cpu_qlen set to 1000 Sep 13 01:20:32.613345 kernel: PTP clock support registered Sep 13 01:20:32.613365 kernel: libata version 3.00 loaded. Sep 13 01:20:32.623345 kernel: ACPI: bus type USB registered Sep 13 01:20:32.644785 kernel: usbcore: registered new interface driver usbfs Sep 13 01:20:32.654343 kernel: usbcore: registered new interface driver hub Sep 13 01:20:32.654360 kernel: usbcore: registered new device driver usb Sep 13 01:20:32.683861 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:20:32.771455 kernel: AVX2 version of gcm_enc/dec engaged. Sep 13 01:20:32.771471 kernel: AES CTR mode by8 optimization enabled Sep 13 01:20:32.771479 kernel: ahci 0000:00:17.0: version 3.0 Sep 13 01:20:32.771571 kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 7 ports 6 Gbps 0x7f impl SATA mode Sep 13 01:20:32.771642 kernel: ahci 0000:00:17.0: flags: 64bit ncq sntf clo only pio slum part ems deso sadm sds apst Sep 13 01:20:32.771705 kernel: scsi host0: ahci Sep 13 01:20:32.771772 kernel: scsi host1: ahci Sep 13 01:20:32.731445 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 01:20:32.806659 kernel: scsi host2: ahci Sep 13 01:20:32.806743 kernel: scsi host3: ahci Sep 13 01:20:32.806807 kernel: scsi host4: ahci Sep 13 01:20:32.731493 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:32.980515 kernel: scsi host5: ahci Sep 13 01:20:32.980616 kernel: scsi host6: ahci Sep 13 01:20:32.980696 kernel: ata1: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516100 irq 127 Sep 13 01:20:32.980705 kernel: ata2: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516180 irq 127 Sep 13 01:20:32.980713 kernel: ata3: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516200 irq 127 Sep 13 01:20:32.980720 kernel: ata4: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516280 irq 127 Sep 13 01:20:32.980727 kernel: ata5: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516300 irq 127 Sep 13 01:20:32.980734 kernel: ata6: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516380 irq 127 Sep 13 01:20:32.980741 kernel: ata7: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516400 irq 127 Sep 13 01:20:32.980748 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Sep 13 01:20:32.980756 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Sep 13 01:20:32.794401 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:20:33.089596 kernel: mlx5_core 0000:01:00.0: firmware version: 14.31.1014 Sep 13 01:20:33.089745 kernel: igb 0000:03:00.0: added PHC on eth0 Sep 13 01:20:33.089821 kernel: mlx5_core 0000:01:00.0: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Sep 13 01:20:33.089888 kernel: igb 0000:03:00.0: Intel(R) Gigabit Ethernet Network Connection Sep 13 01:20:33.089958 kernel: igb 0000:03:00.0: eth0: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:32:0c Sep 13 01:20:33.090024 kernel: igb 0000:03:00.0: eth0: PBA No: 010000-000 Sep 13 01:20:33.090088 kernel: igb 0000:03:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Sep 13 01:20:32.980604 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:20:32.996789 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 13 01:20:33.090094 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 13 01:20:33.090119 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:20:33.090146 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 13 01:20:33.099471 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 13 01:20:33.360817 kernel: igb 0000:04:00.0: added PHC on eth1 Sep 13 01:20:33.360913 kernel: igb 0000:04:00.0: Intel(R) Gigabit Ethernet Network Connection Sep 13 01:20:33.360982 kernel: igb 0000:04:00.0: eth1: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:6a:32:0d Sep 13 01:20:33.361048 kernel: igb 0000:04:00.0: eth1: PBA No: 010000-000 Sep 13 01:20:33.361112 kernel: ata2: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Sep 13 01:20:33.361120 kernel: igb 0000:04:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Sep 13 01:20:33.361183 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361194 kernel: ata1: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Sep 13 01:20:33.361201 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361208 kernel: ata3: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361216 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361223 kernel: ata7: SATA link down (SStatus 0 SControl 300) Sep 13 01:20:33.361230 kernel: ata2.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Sep 13 01:20:33.361237 kernel: ata1.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Sep 13 01:20:33.361244 kernel: mlx5_core 0000:01:00.0: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Sep 13 01:20:33.361311 kernel: ata2.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Sep 13 01:20:33.361319 kernel: mlx5_core 0000:01:00.0: Port module event: module 0, Cable plugged Sep 13 01:20:33.361391 kernel: ata1.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Sep 13 01:20:33.361399 kernel: ata1.00: Features: NCQ-prio Sep 13 01:20:33.226628 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:33.383332 kernel: ata2.00: Features: NCQ-prio Sep 13 01:20:33.396353 kernel: ata1.00: configured for UDMA/133 Sep 13 01:20:33.396396 kernel: ata2.00: configured for UDMA/133 Sep 13 01:20:33.401713 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:20:33.441136 kernel: scsi 0:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Sep 13 01:20:33.441289 kernel: scsi 1:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Sep 13 01:20:33.455469 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Sep 13 01:20:33.455587 kernel: igb 0000:03:00.0 eno1: renamed from eth0 Sep 13 01:20:33.455671 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 1 Sep 13 01:20:33.479611 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 13 01:20:33.503380 kernel: xhci_hcd 0000:00:14.0: hcc params 0x200077c1 hci version 0x110 quirks 0x0000000000009810 Sep 13 01:20:33.515441 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Sep 13 01:20:33.515564 kernel: igb 0000:04:00.0 eno2: renamed from eth1 Sep 13 01:20:33.515654 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 2 Sep 13 01:20:33.555348 kernel: xhci_hcd 0000:00:14.0: Host supports USB 3.1 Enhanced SuperSpeed Sep 13 01:20:33.564344 kernel: hub 1-0:1.0: USB hub found Sep 13 01:20:33.579976 kernel: hub 1-0:1.0: 16 ports detected Sep 13 01:20:33.602343 kernel: hub 2-0:1.0: USB hub found Sep 13 01:20:33.602436 kernel: mlx5_core 0000:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Sep 13 01:20:33.602511 kernel: hub 2-0:1.0: 10 ports detected Sep 13 01:20:33.615345 kernel: mlx5_core 0000:01:00.1: firmware version: 14.31.1014 Sep 13 01:20:33.656304 kernel: mlx5_core 0000:01:00.1: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Sep 13 01:20:33.663068 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:33.794297 kernel: ata2.00: Enabling discard_zeroes_data Sep 13 01:20:33.794311 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:33.794319 kernel: sd 1:0:0:0: [sda] 937703088 512-byte logical blocks: (480 GB/447 GiB) Sep 13 01:20:33.794409 kernel: sd 0:0:0:0: [sdb] 937703088 512-byte logical blocks: (480 GB/447 GiB) Sep 13 01:20:33.794478 kernel: sd 1:0:0:0: [sda] 4096-byte physical blocks Sep 13 01:20:33.794540 kernel: sd 0:0:0:0: [sdb] 4096-byte physical blocks Sep 13 01:20:33.794603 kernel: sd 1:0:0:0: [sda] Write Protect is off Sep 13 01:20:33.794665 kernel: sd 1:0:0:0: [sda] Mode Sense: 00 3a 00 00 Sep 13 01:20:33.794727 kernel: sd 1:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Sep 13 01:20:33.794788 kernel: sd 1:0:0:0: [sda] Preferred minimum I/O size 4096 bytes Sep 13 01:20:33.794849 kernel: sd 0:0:0:0: [sdb] Write Protect is off Sep 13 01:20:33.794908 kernel: ata2.00: Enabling discard_zeroes_data Sep 13 01:20:33.794916 kernel: sd 1:0:0:0: [sda] Attached SCSI disk Sep 13 01:20:33.840846 kernel: usb 1-14: new high-speed USB device number 2 using xhci_hcd Sep 13 01:20:33.840876 kernel: sd 0:0:0:0: [sdb] Mode Sense: 00 3a 00 00 Sep 13 01:20:33.876005 kernel: sd 0:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Sep 13 01:20:33.885399 kernel: sd 0:0:0:0: [sdb] Preferred minimum I/O size 4096 bytes Sep 13 01:20:33.900409 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:33.935438 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 13 01:20:33.935456 kernel: GPT:9289727 != 937703087 Sep 13 01:20:33.950220 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 13 01:20:33.962829 kernel: GPT:9289727 != 937703087 Sep 13 01:20:33.962848 kernel: mlx5_core 0000:01:00.1: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Sep 13 01:20:33.962935 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 13 01:20:33.962944 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:33.967379 kernel: hub 1-14:1.0: USB hub found Sep 13 01:20:33.967473 kernel: hub 1-14:1.0: 4 ports detected Sep 13 01:20:33.990980 kernel: mlx5_core 0000:01:00.1: Port module event: module 1, Cable plugged Sep 13 01:20:33.991069 kernel: sd 0:0:0:0: [sdb] Attached SCSI disk Sep 13 01:20:34.110459 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Micron_5300_MTFDDAK480TDT ROOT. Sep 13 01:20:34.165588 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sdb6 scanned by (udev-worker) (680) Sep 13 01:20:34.165603 kernel: BTRFS: device fsid fa70a3b0-3d47-4508-bba0-9fa4607626aa devid 1 transid 36 /dev/sdb3 scanned by (udev-worker) (544) Sep 13 01:20:34.149084 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Micron_5300_MTFDDAK480TDT EFI-SYSTEM. Sep 13 01:20:34.184632 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Sep 13 01:20:34.192981 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Micron_5300_MTFDDAK480TDT USR-A. Sep 13 01:20:34.221605 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Micron_5300_MTFDDAK480TDT USR-A. Sep 13 01:20:34.306582 kernel: mlx5_core 0000:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Sep 13 01:20:34.306678 kernel: usb 1-14.1: new low-speed USB device number 3 using xhci_hcd Sep 13 01:20:34.306697 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: renamed from eth0 Sep 13 01:20:34.307515 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 13 01:20:34.358394 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: renamed from eth1 Sep 13 01:20:34.358486 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:34.358495 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:34.358539 disk-uuid[731]: Primary Header is updated. Sep 13 01:20:34.358539 disk-uuid[731]: Secondary Entries is updated. Sep 13 01:20:34.358539 disk-uuid[731]: Secondary Header is updated. Sep 13 01:20:34.407454 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:34.407466 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:34.407473 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:34.430345 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:34.453405 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 13 01:20:34.475584 kernel: usbcore: registered new interface driver usbhid Sep 13 01:20:34.475612 kernel: usbhid: USB HID core driver Sep 13 01:20:34.520414 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.0/0003:0557:2419.0001/input/input0 Sep 13 01:20:34.616423 kernel: hid-generic 0003:0557:2419.0001: input,hidraw0: USB HID v1.00 Keyboard [HID 0557:2419] on usb-0000:00:14.0-14.1/input0 Sep 13 01:20:34.616546 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.1/0003:0557:2419.0002/input/input1 Sep 13 01:20:34.650314 kernel: hid-generic 0003:0557:2419.0002: input,hidraw1: USB HID v1.00 Mouse [HID 0557:2419] on usb-0000:00:14.0-14.1/input1 Sep 13 01:20:35.413889 kernel: ata1.00: Enabling discard_zeroes_data Sep 13 01:20:35.433278 disk-uuid[733]: The operation has completed successfully. Sep 13 01:20:35.441450 kernel: sdb: sdb1 sdb2 sdb3 sdb4 sdb6 sdb7 sdb9 Sep 13 01:20:35.475290 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 13 01:20:35.475353 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 13 01:20:35.505630 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 13 01:20:35.532133 sh[752]: Success Sep 13 01:20:35.560383 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 13 01:20:35.610995 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 13 01:20:35.641591 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 13 01:20:35.649656 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 13 01:20:35.705201 kernel: BTRFS info (device dm-0): first mount of filesystem fa70a3b0-3d47-4508-bba0-9fa4607626aa Sep 13 01:20:35.705226 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:35.727003 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 13 01:20:35.746501 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 13 01:20:35.764911 kernel: BTRFS info (device dm-0): using free space tree Sep 13 01:20:35.805375 kernel: BTRFS info (device dm-0): enabling ssd optimizations Sep 13 01:20:35.808108 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 13 01:20:35.816778 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 13 01:20:35.830598 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 13 01:20:35.848752 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 13 01:20:35.892891 kernel: BTRFS info (device sdb6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:35.892912 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:35.911481 kernel: BTRFS info (device sdb6): using free space tree Sep 13 01:20:35.937604 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 13 01:20:35.994621 kernel: BTRFS info (device sdb6): enabling ssd optimizations Sep 13 01:20:35.994638 kernel: BTRFS info (device sdb6): auto enabling async discard Sep 13 01:20:35.994646 kernel: BTRFS info (device sdb6): last unmount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:35.997638 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 13 01:20:36.007627 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 13 01:20:36.008649 systemd-networkd[933]: lo: Link UP Sep 13 01:20:36.008651 systemd-networkd[933]: lo: Gained carrier Sep 13 01:20:36.011122 systemd-networkd[933]: Enumeration completed Sep 13 01:20:36.011866 systemd-networkd[933]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.017684 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 13 01:20:36.041456 systemd-networkd[933]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.044872 systemd[1]: Reached target network.target - Network. Sep 13 01:20:36.069921 systemd-networkd[933]: enp1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.073499 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 13 01:20:36.143864 unknown[937]: fetched base config from "system" Sep 13 01:20:36.141750 ignition[937]: Ignition 2.19.0 Sep 13 01:20:36.143869 unknown[937]: fetched user config from "system" Sep 13 01:20:36.141754 ignition[937]: Stage: fetch-offline Sep 13 01:20:36.145155 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 13 01:20:36.141778 ignition[937]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:36.158652 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 13 01:20:36.141784 ignition[937]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:36.172701 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 13 01:20:36.266593 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Sep 13 01:20:36.141841 ignition[937]: parsed url from cmdline: "" Sep 13 01:20:36.260800 systemd-networkd[933]: enp1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:20:36.141843 ignition[937]: no config URL provided Sep 13 01:20:36.141846 ignition[937]: reading system config file "/usr/lib/ignition/user.ign" Sep 13 01:20:36.141872 ignition[937]: parsing config with SHA512: 26b2c5bf2a30dc9da010c8bdf3f2719492292f4d9470ffc922d9935718c9a83cfef36d87e1f121dc535c90050c9d5379081ae80fca0014dc6a76082ef4a3c8c8 Sep 13 01:20:36.144072 ignition[937]: fetch-offline: fetch-offline passed Sep 13 01:20:36.144075 ignition[937]: POST message to Packet Timeline Sep 13 01:20:36.144077 ignition[937]: POST Status error: resource requires networking Sep 13 01:20:36.144364 ignition[937]: Ignition finished successfully Sep 13 01:20:36.183536 ignition[950]: Ignition 2.19.0 Sep 13 01:20:36.183543 ignition[950]: Stage: kargs Sep 13 01:20:36.183709 ignition[950]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:36.183720 ignition[950]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:36.184596 ignition[950]: kargs: kargs passed Sep 13 01:20:36.184601 ignition[950]: POST message to Packet Timeline Sep 13 01:20:36.184614 ignition[950]: GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:36.185273 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:43970->[::1]:53: read: connection refused Sep 13 01:20:36.385986 ignition[950]: GET https://metadata.packet.net/metadata: attempt #2 Sep 13 01:20:36.387247 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:58372->[::1]:53: read: connection refused Sep 13 01:20:36.520447 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Sep 13 01:20:36.523332 systemd-networkd[933]: eno1: Link UP Sep 13 01:20:36.523478 systemd-networkd[933]: eno2: Link UP Sep 13 01:20:36.523601 systemd-networkd[933]: enp1s0f0np0: Link UP Sep 13 01:20:36.523742 systemd-networkd[933]: enp1s0f0np0: Gained carrier Sep 13 01:20:36.534534 systemd-networkd[933]: enp1s0f1np1: Link UP Sep 13 01:20:36.559448 systemd-networkd[933]: enp1s0f0np0: DHCPv4 address 139.178.94.15/31, gateway 139.178.94.14 acquired from 145.40.83.140 Sep 13 01:20:36.787524 ignition[950]: GET https://metadata.packet.net/metadata: attempt #3 Sep 13 01:20:36.788670 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:48951->[::1]:53: read: connection refused Sep 13 01:20:37.308873 systemd-networkd[933]: enp1s0f1np1: Gained carrier Sep 13 01:20:37.588994 ignition[950]: GET https://metadata.packet.net/metadata: attempt #4 Sep 13 01:20:37.590003 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:57512->[::1]:53: read: connection refused Sep 13 01:20:38.012945 systemd-networkd[933]: enp1s0f0np0: Gained IPv6LL Sep 13 01:20:38.396671 systemd-networkd[933]: enp1s0f1np1: Gained IPv6LL Sep 13 01:20:39.190579 ignition[950]: GET https://metadata.packet.net/metadata: attempt #5 Sep 13 01:20:39.191662 ignition[950]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:40537->[::1]:53: read: connection refused Sep 13 01:20:42.394328 ignition[950]: GET https://metadata.packet.net/metadata: attempt #6 Sep 13 01:20:43.402394 ignition[950]: GET result: OK Sep 13 01:20:43.806538 ignition[950]: Ignition finished successfully Sep 13 01:20:43.810824 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 13 01:20:43.834618 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 13 01:20:43.840733 ignition[969]: Ignition 2.19.0 Sep 13 01:20:43.840738 ignition[969]: Stage: disks Sep 13 01:20:43.840842 ignition[969]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:43.840848 ignition[969]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:43.841362 ignition[969]: disks: disks passed Sep 13 01:20:43.841365 ignition[969]: POST message to Packet Timeline Sep 13 01:20:43.841374 ignition[969]: GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:45.035326 ignition[969]: GET result: OK Sep 13 01:20:45.509175 ignition[969]: Ignition finished successfully Sep 13 01:20:45.512466 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 13 01:20:45.527583 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 13 01:20:45.545605 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 13 01:20:45.556073 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 13 01:20:45.587644 systemd[1]: Reached target sysinit.target - System Initialization. Sep 13 01:20:45.607738 systemd[1]: Reached target basic.target - Basic System. Sep 13 01:20:45.640586 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 13 01:20:45.674920 systemd-fsck[987]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 13 01:20:45.686132 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 13 01:20:45.705676 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 13 01:20:45.808406 kernel: EXT4-fs (sdb9): mounted filesystem 3a3ecd49-b269-4fcb-bb61-e2994e1868ee r/w with ordered data mode. Quota mode: none. Sep 13 01:20:45.808887 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 13 01:20:45.818784 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 13 01:20:45.855563 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 13 01:20:45.864494 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 13 01:20:45.989727 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sdb6 scanned by mount (997) Sep 13 01:20:45.989741 kernel: BTRFS info (device sdb6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:45.989749 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:45.989756 kernel: BTRFS info (device sdb6): using free space tree Sep 13 01:20:45.989763 kernel: BTRFS info (device sdb6): enabling ssd optimizations Sep 13 01:20:45.989770 kernel: BTRFS info (device sdb6): auto enabling async discard Sep 13 01:20:45.904984 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Sep 13 01:20:45.990156 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Sep 13 01:20:46.018603 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 13 01:20:46.018621 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 13 01:20:46.085443 coreos-metadata[999]: Sep 13 01:20:46.065 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Sep 13 01:20:46.038549 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 13 01:20:46.106649 coreos-metadata[1015]: Sep 13 01:20:46.070 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Sep 13 01:20:46.076540 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 13 01:20:46.106578 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 13 01:20:46.152514 initrd-setup-root[1029]: cut: /sysroot/etc/passwd: No such file or directory Sep 13 01:20:46.162589 initrd-setup-root[1036]: cut: /sysroot/etc/group: No such file or directory Sep 13 01:20:46.172557 initrd-setup-root[1043]: cut: /sysroot/etc/shadow: No such file or directory Sep 13 01:20:46.182464 initrd-setup-root[1050]: cut: /sysroot/etc/gshadow: No such file or directory Sep 13 01:20:46.202365 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 13 01:20:46.222579 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 13 01:20:46.258582 kernel: BTRFS info (device sdb6): last unmount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:46.241872 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 13 01:20:46.259145 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 13 01:20:46.289838 ignition[1117]: INFO : Ignition 2.19.0 Sep 13 01:20:46.289838 ignition[1117]: INFO : Stage: mount Sep 13 01:20:46.304450 ignition[1117]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:46.304450 ignition[1117]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:46.304450 ignition[1117]: INFO : mount: mount passed Sep 13 01:20:46.304450 ignition[1117]: INFO : POST message to Packet Timeline Sep 13 01:20:46.304450 ignition[1117]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:46.302085 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 13 01:20:47.161379 coreos-metadata[1015]: Sep 13 01:20:47.161 INFO Fetch successful Sep 13 01:20:47.243448 systemd[1]: flatcar-static-network.service: Deactivated successfully. Sep 13 01:20:47.243506 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Sep 13 01:20:47.276426 coreos-metadata[999]: Sep 13 01:20:47.260 INFO Fetch successful Sep 13 01:20:47.288874 coreos-metadata[999]: Sep 13 01:20:47.288 INFO wrote hostname ci-4081.3.5-n-7cd54b3887 to /sysroot/etc/hostname Sep 13 01:20:47.290007 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 13 01:20:47.353546 ignition[1117]: INFO : GET result: OK Sep 13 01:20:48.080554 ignition[1117]: INFO : Ignition finished successfully Sep 13 01:20:48.083719 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 13 01:20:48.115693 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 13 01:20:48.119827 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 13 01:20:48.186344 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sdb6 scanned by mount (1142) Sep 13 01:20:48.215867 kernel: BTRFS info (device sdb6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:20:48.215883 kernel: BTRFS info (device sdb6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:20:48.233877 kernel: BTRFS info (device sdb6): using free space tree Sep 13 01:20:48.273556 kernel: BTRFS info (device sdb6): enabling ssd optimizations Sep 13 01:20:48.273572 kernel: BTRFS info (device sdb6): auto enabling async discard Sep 13 01:20:48.288098 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 13 01:20:48.315680 ignition[1159]: INFO : Ignition 2.19.0 Sep 13 01:20:48.315680 ignition[1159]: INFO : Stage: files Sep 13 01:20:48.330584 ignition[1159]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:48.330584 ignition[1159]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:48.330584 ignition[1159]: DEBUG : files: compiled without relabeling support, skipping Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 13 01:20:48.330584 ignition[1159]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 13 01:20:48.330584 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Sep 13 01:20:48.319288 unknown[1159]: wrote ssh authorized keys file for user: core Sep 13 01:20:48.466585 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 13 01:20:48.575905 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 13 01:20:48.575905 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:48.608694 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Sep 13 01:20:49.141995 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 13 01:20:49.782539 ignition[1159]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 13 01:20:49.782539 ignition[1159]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 13 01:20:49.811645 ignition[1159]: INFO : files: files passed Sep 13 01:20:49.811645 ignition[1159]: INFO : POST message to Packet Timeline Sep 13 01:20:49.811645 ignition[1159]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:50.704146 ignition[1159]: INFO : GET result: OK Sep 13 01:20:51.142646 ignition[1159]: INFO : Ignition finished successfully Sep 13 01:20:51.146583 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 13 01:20:51.184826 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 13 01:20:51.195067 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 13 01:20:51.215726 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 13 01:20:51.215806 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 13 01:20:51.267612 initrd-setup-root-after-ignition[1195]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:20:51.267612 initrd-setup-root-after-ignition[1195]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:20:51.239118 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 13 01:20:51.327547 initrd-setup-root-after-ignition[1199]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:20:51.259725 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 13 01:20:51.292644 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 13 01:20:51.366946 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 13 01:20:51.367195 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 13 01:20:51.387508 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 13 01:20:51.407605 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 13 01:20:51.428933 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 13 01:20:51.442744 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 13 01:20:51.517586 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 13 01:20:51.545810 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 13 01:20:51.564237 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:20:51.568536 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:20:51.600762 systemd[1]: Stopped target timers.target - Timer Units. Sep 13 01:20:51.619667 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 13 01:20:51.619822 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 13 01:20:51.647039 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 13 01:20:51.669065 systemd[1]: Stopped target basic.target - Basic System. Sep 13 01:20:51.688071 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 13 01:20:51.707080 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 13 01:20:51.727954 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 13 01:20:51.748974 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 13 01:20:51.769068 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 13 01:20:51.789988 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 13 01:20:51.811094 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 13 01:20:51.832059 systemd[1]: Stopped target swap.target - Swaps. Sep 13 01:20:51.850828 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 13 01:20:51.851229 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 13 01:20:51.877198 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:20:51.896982 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:20:51.917827 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 13 01:20:51.918289 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:20:51.940962 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 13 01:20:51.941391 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 13 01:20:51.972927 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 13 01:20:51.973394 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 13 01:20:51.993159 systemd[1]: Stopped target paths.target - Path Units. Sep 13 01:20:52.011813 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 13 01:20:52.012289 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:20:52.032962 systemd[1]: Stopped target slices.target - Slice Units. Sep 13 01:20:52.053066 systemd[1]: Stopped target sockets.target - Socket Units. Sep 13 01:20:52.071916 systemd[1]: iscsid.socket: Deactivated successfully. Sep 13 01:20:52.072219 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 13 01:20:52.093110 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 13 01:20:52.093447 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 13 01:20:52.116152 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 13 01:20:52.116585 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 13 01:20:52.136044 systemd[1]: ignition-files.service: Deactivated successfully. Sep 13 01:20:52.248756 ignition[1221]: INFO : Ignition 2.19.0 Sep 13 01:20:52.248756 ignition[1221]: INFO : Stage: umount Sep 13 01:20:52.248756 ignition[1221]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:20:52.248756 ignition[1221]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Sep 13 01:20:52.248756 ignition[1221]: INFO : umount: umount passed Sep 13 01:20:52.248756 ignition[1221]: INFO : POST message to Packet Timeline Sep 13 01:20:52.248756 ignition[1221]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Sep 13 01:20:52.136447 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 13 01:20:52.154030 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Sep 13 01:20:52.154446 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 13 01:20:52.182620 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 13 01:20:52.200621 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 13 01:20:52.200741 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:20:52.239708 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 13 01:20:52.240644 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 13 01:20:52.240859 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:20:52.266919 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 13 01:20:52.267276 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 13 01:20:52.306581 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 13 01:20:52.306926 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 13 01:20:52.306970 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 13 01:20:52.329618 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 13 01:20:52.329682 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 13 01:20:53.291066 ignition[1221]: INFO : GET result: OK Sep 13 01:20:54.076937 ignition[1221]: INFO : Ignition finished successfully Sep 13 01:20:54.080070 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 13 01:20:54.080391 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 13 01:20:54.096649 systemd[1]: Stopped target network.target - Network. Sep 13 01:20:54.112581 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 13 01:20:54.112750 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 13 01:20:54.130670 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 13 01:20:54.130818 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 13 01:20:54.148637 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 13 01:20:54.148760 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 13 01:20:54.166866 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 13 01:20:54.167038 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 13 01:20:54.185841 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 13 01:20:54.186036 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 13 01:20:54.205256 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 13 01:20:54.215497 systemd-networkd[933]: enp1s0f0np0: DHCPv6 lease lost Sep 13 01:20:54.224757 systemd-networkd[933]: enp1s0f1np1: DHCPv6 lease lost Sep 13 01:20:54.225125 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 13 01:20:54.245465 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 13 01:20:54.245757 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 13 01:20:54.265524 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 13 01:20:54.265823 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 13 01:20:54.287241 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 13 01:20:54.287474 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:20:54.317633 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 13 01:20:54.325701 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 13 01:20:54.325845 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 13 01:20:54.354666 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 01:20:54.354798 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:20:54.372734 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 13 01:20:54.372891 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 13 01:20:54.390758 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 13 01:20:54.390926 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:20:54.411980 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:20:54.432648 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 13 01:20:54.433031 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:20:54.463751 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 13 01:20:54.463788 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 13 01:20:54.471695 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 13 01:20:54.471718 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:20:54.499642 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 13 01:20:54.499712 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 13 01:20:54.540554 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 13 01:20:54.540802 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 13 01:20:54.579465 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 01:20:54.579706 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:20:54.622437 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 13 01:20:54.650406 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 13 01:20:54.865596 systemd-journald[265]: Received SIGTERM from PID 1 (systemd). Sep 13 01:20:54.650448 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:20:54.669489 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 01:20:54.669571 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:54.692696 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 13 01:20:54.692938 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 13 01:20:54.724172 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 13 01:20:54.724453 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 13 01:20:54.744606 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 13 01:20:54.780757 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 13 01:20:54.801433 systemd[1]: Switching root. Sep 13 01:20:54.959527 systemd-journald[265]: Journal stopped Sep 13 01:20:57.610164 kernel: SELinux: policy capability network_peer_controls=1 Sep 13 01:20:57.610179 kernel: SELinux: policy capability open_perms=1 Sep 13 01:20:57.610186 kernel: SELinux: policy capability extended_socket_class=1 Sep 13 01:20:57.610193 kernel: SELinux: policy capability always_check_network=0 Sep 13 01:20:57.610198 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 13 01:20:57.610204 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 13 01:20:57.610210 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 13 01:20:57.610216 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 13 01:20:57.610222 kernel: audit: type=1403 audit(1757726455.196:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 13 01:20:57.610228 systemd[1]: Successfully loaded SELinux policy in 172.108ms. Sep 13 01:20:57.610237 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 7.534ms. Sep 13 01:20:57.610244 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 13 01:20:57.610250 systemd[1]: Detected architecture x86-64. Sep 13 01:20:57.610256 systemd[1]: Detected first boot. Sep 13 01:20:57.610263 systemd[1]: Hostname set to . Sep 13 01:20:57.610271 systemd[1]: Initializing machine ID from random generator. Sep 13 01:20:57.610278 zram_generator::config[1273]: No configuration found. Sep 13 01:20:57.610285 systemd[1]: Populated /etc with preset unit settings. Sep 13 01:20:57.610291 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 13 01:20:57.610297 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 13 01:20:57.610304 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 13 01:20:57.610311 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 13 01:20:57.610318 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 13 01:20:57.610325 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 13 01:20:57.610332 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 13 01:20:57.610343 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 13 01:20:57.610350 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 13 01:20:57.610357 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 13 01:20:57.610364 systemd[1]: Created slice user.slice - User and Session Slice. Sep 13 01:20:57.610372 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:20:57.610379 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:20:57.610385 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 13 01:20:57.610392 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 13 01:20:57.610399 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 13 01:20:57.610405 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 13 01:20:57.610412 systemd[1]: Expecting device dev-ttyS1.device - /dev/ttyS1... Sep 13 01:20:57.610419 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:20:57.610426 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 13 01:20:57.610433 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 13 01:20:57.610440 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 13 01:20:57.610448 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 13 01:20:57.610455 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:20:57.610462 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 13 01:20:57.610469 systemd[1]: Reached target slices.target - Slice Units. Sep 13 01:20:57.610477 systemd[1]: Reached target swap.target - Swaps. Sep 13 01:20:57.610484 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 13 01:20:57.610491 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 13 01:20:57.610498 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:20:57.610507 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 13 01:20:57.610514 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:20:57.610522 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 13 01:20:57.610529 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 13 01:20:57.610536 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 13 01:20:57.610543 systemd[1]: Mounting media.mount - External Media Directory... Sep 13 01:20:57.610550 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:20:57.610558 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 13 01:20:57.610564 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 13 01:20:57.610572 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 13 01:20:57.610580 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 13 01:20:57.610587 systemd[1]: Reached target machines.target - Containers. Sep 13 01:20:57.610594 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 13 01:20:57.610601 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 13 01:20:57.610608 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 13 01:20:57.610615 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 13 01:20:57.610622 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 13 01:20:57.610629 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 13 01:20:57.610637 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 13 01:20:57.610644 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 13 01:20:57.610651 kernel: ACPI: bus type drm_connector registered Sep 13 01:20:57.610658 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 13 01:20:57.610664 kernel: fuse: init (API version 7.39) Sep 13 01:20:57.610671 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 13 01:20:57.610678 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 13 01:20:57.610685 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 13 01:20:57.610693 kernel: loop: module loaded Sep 13 01:20:57.610699 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 13 01:20:57.610706 systemd[1]: Stopped systemd-fsck-usr.service. Sep 13 01:20:57.610713 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 13 01:20:57.610729 systemd-journald[1376]: Collecting audit messages is disabled. Sep 13 01:20:57.610744 systemd-journald[1376]: Journal started Sep 13 01:20:57.610759 systemd-journald[1376]: Runtime Journal (/run/log/journal/ac4d2f7248264ec7a1a82fe342650341) is 8.0M, max 639.9M, 631.9M free. Sep 13 01:20:55.740182 systemd[1]: Queued start job for default target multi-user.target. Sep 13 01:20:55.760185 systemd[1]: Unnecessary job was removed for dev-sdb6.device - /dev/sdb6. Sep 13 01:20:55.760451 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 13 01:20:57.639371 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 13 01:20:57.675344 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 13 01:20:57.708388 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 13 01:20:57.741374 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 13 01:20:57.774772 systemd[1]: verity-setup.service: Deactivated successfully. Sep 13 01:20:57.774802 systemd[1]: Stopped verity-setup.service. Sep 13 01:20:57.837390 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:20:57.858534 systemd[1]: Started systemd-journald.service - Journal Service. Sep 13 01:20:57.868930 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 13 01:20:57.879626 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 13 01:20:57.889610 systemd[1]: Mounted media.mount - External Media Directory. Sep 13 01:20:57.899593 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 13 01:20:57.909608 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 13 01:20:57.919572 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 13 01:20:57.929684 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 13 01:20:57.940733 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:20:57.951917 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 13 01:20:57.952140 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 13 01:20:57.964210 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 01:20:57.964660 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 13 01:20:57.976236 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 01:20:57.976626 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 13 01:20:57.987232 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 01:20:57.987626 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 13 01:20:57.999231 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 13 01:20:57.999614 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 13 01:20:58.010259 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 01:20:58.010648 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 13 01:20:58.021268 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 13 01:20:58.032186 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 13 01:20:58.044190 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 13 01:20:58.056184 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:20:58.089850 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 13 01:20:58.116577 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 13 01:20:58.127171 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 13 01:20:58.136516 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 13 01:20:58.136541 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 13 01:20:58.147393 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 13 01:20:58.166609 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 13 01:20:58.179877 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 13 01:20:58.189812 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 13 01:20:58.193254 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 13 01:20:58.202970 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 13 01:20:58.213460 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 01:20:58.214210 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 13 01:20:58.216908 systemd-journald[1376]: Time spent on flushing to /var/log/journal/ac4d2f7248264ec7a1a82fe342650341 is 14.791ms for 1367 entries. Sep 13 01:20:58.216908 systemd-journald[1376]: System Journal (/var/log/journal/ac4d2f7248264ec7a1a82fe342650341) is 8.0M, max 195.6M, 187.6M free. Sep 13 01:20:58.262960 systemd-journald[1376]: Received client request to flush runtime journal. Sep 13 01:20:58.232496 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 13 01:20:58.233125 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 13 01:20:58.249176 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 13 01:20:58.261264 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 13 01:20:58.277093 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 13 01:20:58.286381 kernel: loop0: detected capacity change from 0 to 140768 Sep 13 01:20:58.287101 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 13 01:20:58.312033 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 13 01:20:58.325345 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 13 01:20:58.335616 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 13 01:20:58.346650 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 13 01:20:58.357580 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 13 01:20:58.368509 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:20:58.387376 kernel: loop1: detected capacity change from 0 to 142488 Sep 13 01:20:58.396584 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 13 01:20:58.409269 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 13 01:20:58.434593 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 13 01:20:58.446095 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 13 01:20:58.461118 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 13 01:20:58.461668 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 13 01:20:58.474409 kernel: loop2: detected capacity change from 0 to 224512 Sep 13 01:20:58.484852 udevadm[1412]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Sep 13 01:20:58.509732 systemd-tmpfiles[1425]: ACLs are not supported, ignoring. Sep 13 01:20:58.509742 systemd-tmpfiles[1425]: ACLs are not supported, ignoring. Sep 13 01:20:58.512332 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:20:58.544391 kernel: loop3: detected capacity change from 0 to 8 Sep 13 01:20:58.577185 ldconfig[1402]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 13 01:20:58.578610 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 13 01:20:58.604513 kernel: loop4: detected capacity change from 0 to 140768 Sep 13 01:20:58.642330 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 13 01:20:58.651379 kernel: loop5: detected capacity change from 0 to 142488 Sep 13 01:20:58.684397 kernel: loop6: detected capacity change from 0 to 224512 Sep 13 01:20:58.685524 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:20:58.697888 systemd-udevd[1434]: Using default interface naming scheme 'v255'. Sep 13 01:20:58.713231 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:20:58.715346 kernel: loop7: detected capacity change from 0 to 8 Sep 13 01:20:58.715520 (sd-merge)[1431]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-packet'. Sep 13 01:20:58.715758 (sd-merge)[1431]: Merged extensions into '/usr'. Sep 13 01:20:58.734080 systemd[1]: Reloading requested from client PID 1407 ('systemd-sysext') (unit systemd-sysext.service)... Sep 13 01:20:58.734087 systemd[1]: Reloading... Sep 13 01:20:58.749223 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input2 Sep 13 01:20:58.749271 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 36 scanned by (udev-worker) (1456) Sep 13 01:20:58.749307 kernel: ACPI: button: Sleep Button [SLPB] Sep 13 01:20:58.784479 zram_generator::config[1535]: No configuration found. Sep 13 01:20:58.784568 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 13 01:20:58.847376 kernel: mousedev: PS/2 mouse device common for all mice Sep 13 01:20:58.847428 kernel: ACPI: button: Power Button [PWRF] Sep 13 01:20:58.864722 kernel: IPMI message handler: version 39.2 Sep 13 01:20:58.896291 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:20:58.897345 kernel: ipmi device interface Sep 13 01:20:58.924876 kernel: i801_smbus 0000:00:1f.4: SPD Write Disable is set Sep 13 01:20:58.925146 kernel: i801_smbus 0000:00:1f.4: SMBus using PCI interrupt Sep 13 01:20:58.954347 kernel: i2c i2c-0: 2/4 memory slots populated (from DMI) Sep 13 01:20:58.957345 kernel: ipmi_si: IPMI System Interface driver Sep 13 01:20:58.957363 kernel: mei_me 0000:00:16.0: Device doesn't have valid ME Interface Sep 13 01:20:58.957461 kernel: mei_me 0000:00:16.4: Device doesn't have valid ME Interface Sep 13 01:20:58.968129 kernel: ipmi_si dmi-ipmi-si.0: ipmi_platform: probing via SMBIOS Sep 13 01:20:58.976034 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Sep 13 01:20:59.032470 kernel: ipmi_platform: ipmi_si: SMBIOS: io 0xca2 regsize 1 spacing 1 irq 0 Sep 13 01:20:59.032501 kernel: ipmi_si: Adding SMBIOS-specified kcs state machine Sep 13 01:20:59.032515 kernel: ipmi_si IPI0001:00: ipmi_platform: probing via ACPI Sep 13 01:20:59.055050 kernel: iTCO_vendor_support: vendor-support=0 Sep 13 01:20:59.055078 kernel: ipmi_si IPI0001:00: ipmi_platform: [io 0x0ca2] regsize 1 spacing 1 irq 0 Sep 13 01:20:59.055070 systemd[1]: Condition check resulted in dev-ttyS1.device - /dev/ttyS1 being skipped. Sep 13 01:20:59.055294 systemd[1]: Reloading finished in 321 ms. Sep 13 01:20:59.110347 kernel: ipmi_si dmi-ipmi-si.0: Removing SMBIOS-specified kcs state machine in favor of ACPI Sep 13 01:20:59.134410 kernel: ipmi_si: Adding ACPI-specified kcs state machine Sep 13 01:20:59.154503 kernel: ipmi_si: Trying ACPI-specified kcs state machine at i/o address 0xca2, slave address 0x20, irq 0 Sep 13 01:20:59.214003 kernel: iTCO_wdt iTCO_wdt: Found a Intel PCH TCO device (Version=6, TCOBASE=0x0400) Sep 13 01:20:59.214176 kernel: iTCO_wdt iTCO_wdt: initialized. heartbeat=30 sec (nowayout=0) Sep 13 01:20:59.224343 kernel: ipmi_si IPI0001:00: The BMC does not support clearing the recv irq bit, compensating, but the BMC needs to be fixed. Sep 13 01:20:59.267036 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 13 01:20:59.279386 kernel: intel_rapl_common: Found RAPL domain package Sep 13 01:20:59.279412 kernel: ipmi_si IPI0001:00: IPMI message handler: Found new BMC (man_id: 0x002a7c, prod_id: 0x1b0f, dev_id: 0x20) Sep 13 01:20:59.279517 kernel: intel_rapl_common: Found RAPL domain core Sep 13 01:20:59.279530 kernel: intel_rapl_common: Found RAPL domain dram Sep 13 01:20:59.354602 systemd[1]: Starting ensure-sysext.service... Sep 13 01:20:59.362046 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 13 01:20:59.384818 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 13 01:20:59.398881 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 13 01:20:59.409324 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:20:59.409406 kernel: ipmi_si IPI0001:00: IPMI kcs interface initialized Sep 13 01:20:59.410858 systemd[1]: Reloading requested from client PID 1604 ('systemctl') (unit ensure-sysext.service)... Sep 13 01:20:59.410869 systemd[1]: Reloading... Sep 13 01:20:59.427345 kernel: ipmi_ssif: IPMI SSIF Interface driver Sep 13 01:20:59.433614 systemd-tmpfiles[1609]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 13 01:20:59.433988 systemd-tmpfiles[1609]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 13 01:20:59.434849 systemd-tmpfiles[1609]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 13 01:20:59.435059 systemd-tmpfiles[1609]: ACLs are not supported, ignoring. Sep 13 01:20:59.435119 systemd-tmpfiles[1609]: ACLs are not supported, ignoring. Sep 13 01:20:59.437250 systemd-tmpfiles[1609]: Detected autofs mount point /boot during canonicalization of boot. Sep 13 01:20:59.437258 systemd-tmpfiles[1609]: Skipping /boot Sep 13 01:20:59.441659 systemd-tmpfiles[1609]: Detected autofs mount point /boot during canonicalization of boot. Sep 13 01:20:59.441664 systemd-tmpfiles[1609]: Skipping /boot Sep 13 01:20:59.452425 zram_generator::config[1641]: No configuration found. Sep 13 01:20:59.507404 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:20:59.561996 systemd[1]: Reloading finished in 150 ms. Sep 13 01:20:59.577934 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 13 01:20:59.598594 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 13 01:20:59.609594 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:20:59.620562 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:20:59.643650 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 13 01:20:59.654800 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 13 01:20:59.662189 augenrules[1724]: No rules Sep 13 01:20:59.666206 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 13 01:20:59.678413 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 13 01:20:59.685492 lvm[1729]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 01:20:59.690790 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 13 01:20:59.702137 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 13 01:20:59.715400 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 13 01:20:59.725990 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 13 01:20:59.735680 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 13 01:20:59.750503 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 13 01:20:59.760696 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 13 01:20:59.771690 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 13 01:20:59.784422 systemd-networkd[1608]: lo: Link UP Sep 13 01:20:59.784428 systemd-networkd[1608]: lo: Gained carrier Sep 13 01:20:59.786548 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:20:59.787146 systemd-networkd[1608]: bond0: netdev ready Sep 13 01:20:59.788086 systemd-networkd[1608]: Enumeration completed Sep 13 01:20:59.796513 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:20:59.796631 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 13 01:20:59.797383 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 13 01:20:59.800456 systemd-networkd[1608]: enp1s0f0np0: Configuring with /etc/systemd/network/10-0c:42:a1:15:ae:cc.network. Sep 13 01:20:59.809092 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 13 01:20:59.810966 lvm[1749]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 01:20:59.819062 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 13 01:20:59.826494 systemd-resolved[1731]: Positive Trust Anchors: Sep 13 01:20:59.826499 systemd-resolved[1731]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 01:20:59.826523 systemd-resolved[1731]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 13 01:20:59.829459 systemd-resolved[1731]: Using system hostname 'ci-4081.3.5-n-7cd54b3887'. Sep 13 01:20:59.832043 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 13 01:20:59.841421 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 13 01:20:59.842118 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 13 01:20:59.851381 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 01:20:59.851470 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:20:59.852043 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 13 01:20:59.861887 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 13 01:20:59.872606 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 13 01:20:59.883621 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 01:20:59.883692 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 13 01:20:59.894655 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 01:20:59.894733 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 13 01:20:59.905652 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 01:20:59.905723 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 13 01:20:59.915638 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 13 01:20:59.928380 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:20:59.928509 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 13 01:20:59.937555 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 13 01:20:59.947993 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 13 01:20:59.959017 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 13 01:20:59.968477 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 13 01:20:59.969295 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 13 01:20:59.980450 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 01:20:59.980510 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:20:59.981146 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 01:20:59.981219 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 13 01:20:59.992663 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 01:20:59.992734 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 13 01:21:00.003658 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 01:21:00.003726 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 13 01:21:00.015663 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:21:00.015790 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 13 01:21:00.026493 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 13 01:21:00.037001 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 13 01:21:00.046988 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 13 01:21:00.057901 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 13 01:21:00.067473 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 13 01:21:00.067554 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 01:21:00.067610 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:21:00.068208 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 01:21:00.068285 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 13 01:21:00.079633 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 01:21:00.079708 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 13 01:21:00.089637 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 01:21:00.089708 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 13 01:21:00.100591 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 01:21:00.100664 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 13 01:21:00.111280 systemd[1]: Finished ensure-sysext.service. Sep 13 01:21:00.121814 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 01:21:00.121847 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 13 01:21:00.142529 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 13 01:21:00.177731 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 13 01:21:00.188456 systemd[1]: Reached target time-set.target - System Time Set. Sep 13 01:21:00.590373 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Sep 13 01:21:00.613396 kernel: bond0: (slave enp1s0f0np0): Enslaving as a backup interface with an up link Sep 13 01:21:00.614841 systemd-networkd[1608]: enp1s0f1np1: Configuring with /etc/systemd/network/10-0c:42:a1:15:ae:cd.network. Sep 13 01:21:00.615745 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 13 01:21:00.626499 systemd[1]: Reached target network.target - Network. Sep 13 01:21:00.634510 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:21:00.645428 systemd[1]: Reached target sysinit.target - System Initialization. Sep 13 01:21:00.655481 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 13 01:21:00.666523 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 13 01:21:00.677699 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 13 01:21:00.687691 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 13 01:21:00.698593 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 13 01:21:00.709570 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 13 01:21:00.709645 systemd[1]: Reached target paths.target - Path Units. Sep 13 01:21:00.717579 systemd[1]: Reached target timers.target - Timer Units. Sep 13 01:21:00.728009 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 13 01:21:00.738870 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 13 01:21:00.752286 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 13 01:21:00.762111 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 13 01:21:00.771725 systemd[1]: Reached target sockets.target - Socket Units. Sep 13 01:21:00.781603 systemd[1]: Reached target basic.target - Basic System. Sep 13 01:21:00.789667 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 13 01:21:00.789747 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 13 01:21:00.809512 systemd[1]: Starting containerd.service - containerd container runtime... Sep 13 01:21:00.826382 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Sep 13 01:21:00.827204 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Sep 13 01:21:00.848391 kernel: bond0: (slave enp1s0f1np1): Enslaving as a backup interface with an up link Sep 13 01:21:00.848979 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 13 01:21:00.851512 systemd-networkd[1608]: bond0: Configuring with /etc/systemd/network/05-bond0.network. Sep 13 01:21:00.851858 systemd-networkd[1608]: enp1s0f0np0: Link UP Sep 13 01:21:00.852018 systemd-networkd[1608]: enp1s0f0np0: Gained carrier Sep 13 01:21:00.868012 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 13 01:21:00.872059 coreos-metadata[1782]: Sep 13 01:21:00.872 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Sep 13 01:21:00.873318 coreos-metadata[1782]: Sep 13 01:21:00.873 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Sep 13 01:21:00.873368 kernel: bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond Sep 13 01:21:00.873636 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 13 01:21:00.877729 systemd-networkd[1608]: enp1s0f1np1: Reconfiguring with /etc/systemd/network/10-0c:42:a1:15:ae:cc.network. Sep 13 01:21:00.877873 systemd-networkd[1608]: enp1s0f1np1: Link UP Sep 13 01:21:00.878026 systemd-networkd[1608]: enp1s0f1np1: Gained carrier Sep 13 01:21:00.885131 jq[1786]: false Sep 13 01:21:00.890482 systemd-networkd[1608]: bond0: Link UP Sep 13 01:21:00.890659 systemd-networkd[1608]: bond0: Gained carrier Sep 13 01:21:00.890763 systemd-timesyncd[1777]: Network configuration changed, trying to establish connection. Sep 13 01:21:00.892419 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 13 01:21:00.893225 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 13 01:21:00.897040 dbus-daemon[1783]: [system] SELinux support is enabled Sep 13 01:21:00.901014 extend-filesystems[1788]: Found loop4 Sep 13 01:21:00.901014 extend-filesystems[1788]: Found loop5 Sep 13 01:21:00.947277 kernel: EXT4-fs (sdb9): resizing filesystem from 553472 to 116605649 blocks Sep 13 01:21:00.947306 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 36 scanned by (udev-worker) (1498) Sep 13 01:21:00.903191 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 13 01:21:00.947358 extend-filesystems[1788]: Found loop6 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found loop7 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sda Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb1 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb2 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb3 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found usr Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb4 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb6 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb7 Sep 13 01:21:00.947358 extend-filesystems[1788]: Found sdb9 Sep 13 01:21:00.947358 extend-filesystems[1788]: Checking size of /dev/sdb9 Sep 13 01:21:00.947358 extend-filesystems[1788]: Resized partition /dev/sdb9 Sep 13 01:21:01.150455 kernel: bond0: (slave enp1s0f0np0): link status definitely up, 25000 Mbps full duplex Sep 13 01:21:01.150475 kernel: bond0: active interface up! Sep 13 01:21:01.150487 kernel: bond0: (slave enp1s0f1np1): link status definitely up, 25000 Mbps full duplex Sep 13 01:21:01.001488 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 13 01:21:01.150601 extend-filesystems[1797]: resize2fs 1.47.1 (20-May-2024) Sep 13 01:21:01.022069 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 13 01:21:01.060471 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 13 01:21:01.075355 systemd[1]: Starting tcsd.service - TCG Core Services Daemon... Sep 13 01:21:01.083789 systemd-logind[1809]: Watching system buttons on /dev/input/event3 (Power Button) Sep 13 01:21:01.083803 systemd-logind[1809]: Watching system buttons on /dev/input/event2 (Sleep Button) Sep 13 01:21:01.168949 update_engine[1814]: I20250913 01:21:01.131139 1814 main.cc:92] Flatcar Update Engine starting Sep 13 01:21:01.168949 update_engine[1814]: I20250913 01:21:01.131933 1814 update_check_scheduler.cc:74] Next update check in 11m15s Sep 13 01:21:01.083814 systemd-logind[1809]: Watching system buttons on /dev/input/event0 (HID 0557:2419) Sep 13 01:21:01.169108 jq[1815]: true Sep 13 01:21:01.084015 systemd-logind[1809]: New seat seat0. Sep 13 01:21:01.098674 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 13 01:21:01.099090 systemd[1]: Starting update-engine.service - Update Engine... Sep 13 01:21:01.112047 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 13 01:21:01.123780 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 13 01:21:01.161554 systemd[1]: Started systemd-logind.service - User Login Management. Sep 13 01:21:01.184530 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 13 01:21:01.184623 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 13 01:21:01.184796 systemd[1]: motdgen.service: Deactivated successfully. Sep 13 01:21:01.184884 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 13 01:21:01.194831 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 13 01:21:01.194915 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 13 01:21:01.208254 (ntainerd)[1819]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 13 01:21:01.209718 jq[1818]: true Sep 13 01:21:01.211319 dbus-daemon[1783]: [system] Successfully activated service 'org.freedesktop.systemd1' Sep 13 01:21:01.213915 tar[1817]: linux-amd64/LICENSE Sep 13 01:21:01.214048 tar[1817]: linux-amd64/helm Sep 13 01:21:01.215482 systemd[1]: tcsd.service: Skipped due to 'exec-condition'. Sep 13 01:21:01.215579 systemd[1]: Condition check resulted in tcsd.service - TCG Core Services Daemon being skipped. Sep 13 01:21:01.219672 systemd[1]: Started update-engine.service - Update Engine. Sep 13 01:21:01.231351 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 13 01:21:01.231459 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 13 01:21:01.242419 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 13 01:21:01.242493 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 13 01:21:01.270322 bash[1847]: Updated "/home/core/.ssh/authorized_keys" Sep 13 01:21:01.271508 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 13 01:21:01.283234 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 13 01:21:01.295983 systemd[1]: Starting sshkeys.service... Sep 13 01:21:01.296829 locksmithd[1848]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 13 01:21:01.307476 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Sep 13 01:21:01.319070 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Sep 13 01:21:01.341225 coreos-metadata[1856]: Sep 13 01:21:01.341 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Sep 13 01:21:01.385425 containerd[1819]: time="2025-09-13T01:21:01.385378301Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Sep 13 01:21:01.397897 containerd[1819]: time="2025-09-13T01:21:01.397877448Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398708 containerd[1819]: time="2025-09-13T01:21:01.398690327Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.106-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398708 containerd[1819]: time="2025-09-13T01:21:01.398706573Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 13 01:21:01.398753 containerd[1819]: time="2025-09-13T01:21:01.398715851Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 13 01:21:01.398808 containerd[1819]: time="2025-09-13T01:21:01.398799215Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 13 01:21:01.398830 containerd[1819]: time="2025-09-13T01:21:01.398810202Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398855 containerd[1819]: time="2025-09-13T01:21:01.398843082Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398877 containerd[1819]: time="2025-09-13T01:21:01.398855225Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398958 containerd[1819]: time="2025-09-13T01:21:01.398947442Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398979 containerd[1819]: time="2025-09-13T01:21:01.398957119Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398979 containerd[1819]: time="2025-09-13T01:21:01.398964622Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:21:01.398979 containerd[1819]: time="2025-09-13T01:21:01.398970976Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 13 01:21:01.399031 containerd[1819]: time="2025-09-13T01:21:01.399012984Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:21:01.399141 containerd[1819]: time="2025-09-13T01:21:01.399132685Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:21:01.399196 containerd[1819]: time="2025-09-13T01:21:01.399187247Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:21:01.399220 containerd[1819]: time="2025-09-13T01:21:01.399195929Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 13 01:21:01.399246 containerd[1819]: time="2025-09-13T01:21:01.399238285Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 13 01:21:01.399274 containerd[1819]: time="2025-09-13T01:21:01.399267560Z" level=info msg="metadata content store policy set" policy=shared Sep 13 01:21:01.413550 containerd[1819]: time="2025-09-13T01:21:01.413519026Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 13 01:21:01.413550 containerd[1819]: time="2025-09-13T01:21:01.413542988Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 13 01:21:01.413603 containerd[1819]: time="2025-09-13T01:21:01.413553478Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 13 01:21:01.413603 containerd[1819]: time="2025-09-13T01:21:01.413562583Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 13 01:21:01.413603 containerd[1819]: time="2025-09-13T01:21:01.413570498Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 13 01:21:01.413658 containerd[1819]: time="2025-09-13T01:21:01.413640059Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 13 01:21:01.413950 containerd[1819]: time="2025-09-13T01:21:01.413782892Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 13 01:21:01.414167 containerd[1819]: time="2025-09-13T01:21:01.414150199Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 13 01:21:01.414250 containerd[1819]: time="2025-09-13T01:21:01.414238805Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 13 01:21:01.414270 containerd[1819]: time="2025-09-13T01:21:01.414256797Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 13 01:21:01.414270 containerd[1819]: time="2025-09-13T01:21:01.414267279Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414302 containerd[1819]: time="2025-09-13T01:21:01.414278532Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414302 containerd[1819]: time="2025-09-13T01:21:01.414286749Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414302 containerd[1819]: time="2025-09-13T01:21:01.414294705Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414349 containerd[1819]: time="2025-09-13T01:21:01.414302882Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414349 containerd[1819]: time="2025-09-13T01:21:01.414310795Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414349 containerd[1819]: time="2025-09-13T01:21:01.414318057Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414349 containerd[1819]: time="2025-09-13T01:21:01.414324303Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 13 01:21:01.414349 containerd[1819]: time="2025-09-13T01:21:01.414335729Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414350039Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414357594Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414365090Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414372305Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414380069Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414386870Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414394004Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414401242Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414411269Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414423 containerd[1819]: time="2025-09-13T01:21:01.414418247Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414425873Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414433164Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414441958Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414455194Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414462470Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414468739Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414492440Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414502253Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414508781Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414515944Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414521677Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414531265Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414537926Z" level=info msg="NRI interface is disabled by configuration." Sep 13 01:21:01.414567 containerd[1819]: time="2025-09-13T01:21:01.414544217Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 13 01:21:01.414762 containerd[1819]: time="2025-09-13T01:21:01.414701713Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 13 01:21:01.414762 containerd[1819]: time="2025-09-13T01:21:01.414737011Z" level=info msg="Connect containerd service" Sep 13 01:21:01.414762 containerd[1819]: time="2025-09-13T01:21:01.414755211Z" level=info msg="using legacy CRI server" Sep 13 01:21:01.414762 containerd[1819]: time="2025-09-13T01:21:01.414759625Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 13 01:21:01.414887 containerd[1819]: time="2025-09-13T01:21:01.414809306Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 13 01:21:01.415117 containerd[1819]: time="2025-09-13T01:21:01.415106022Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 01:21:01.415241 containerd[1819]: time="2025-09-13T01:21:01.415221347Z" level=info msg="Start subscribing containerd event" Sep 13 01:21:01.415262 containerd[1819]: time="2025-09-13T01:21:01.415251530Z" level=info msg="Start recovering state" Sep 13 01:21:01.415279 containerd[1819]: time="2025-09-13T01:21:01.415270727Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 13 01:21:01.415301 containerd[1819]: time="2025-09-13T01:21:01.415294018Z" level=info msg="Start event monitor" Sep 13 01:21:01.415319 containerd[1819]: time="2025-09-13T01:21:01.415304185Z" level=info msg="Start snapshots syncer" Sep 13 01:21:01.415319 containerd[1819]: time="2025-09-13T01:21:01.415310247Z" level=info msg="Start cni network conf syncer for default" Sep 13 01:21:01.415319 containerd[1819]: time="2025-09-13T01:21:01.415314329Z" level=info msg="Start streaming server" Sep 13 01:21:01.415373 containerd[1819]: time="2025-09-13T01:21:01.415296826Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 13 01:21:01.415401 containerd[1819]: time="2025-09-13T01:21:01.415393429Z" level=info msg="containerd successfully booted in 0.030455s" Sep 13 01:21:01.415452 systemd[1]: Started containerd.service - containerd container runtime. Sep 13 01:21:01.438393 sshd_keygen[1812]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 13 01:21:01.451320 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 13 01:21:01.480393 kernel: EXT4-fs (sdb9): resized filesystem to 116605649 Sep 13 01:21:01.485584 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 13 01:21:01.493693 systemd[1]: issuegen.service: Deactivated successfully. Sep 13 01:21:01.493783 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 13 01:21:01.502065 extend-filesystems[1797]: Filesystem at /dev/sdb9 is mounted on /; on-line resizing required Sep 13 01:21:01.502065 extend-filesystems[1797]: old_desc_blocks = 1, new_desc_blocks = 56 Sep 13 01:21:01.502065 extend-filesystems[1797]: The filesystem on /dev/sdb9 is now 116605649 (4k) blocks long. Sep 13 01:21:01.532558 extend-filesystems[1788]: Resized filesystem in /dev/sdb9 Sep 13 01:21:01.551383 tar[1817]: linux-amd64/README.md Sep 13 01:21:01.503694 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 13 01:21:01.503781 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 13 01:21:01.553224 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 13 01:21:01.562663 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 13 01:21:01.572843 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 13 01:21:01.594576 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 13 01:21:01.603044 systemd[1]: Started serial-getty@ttyS1.service - Serial Getty on ttyS1. Sep 13 01:21:01.612503 systemd[1]: Reached target getty.target - Login Prompts. Sep 13 01:21:01.873509 coreos-metadata[1782]: Sep 13 01:21:01.873 INFO Fetching https://metadata.packet.net/metadata: Attempt #2 Sep 13 01:21:02.524692 systemd-networkd[1608]: bond0: Gained IPv6LL Sep 13 01:21:02.525926 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 13 01:21:02.538190 systemd[1]: Reached target network-online.target - Network is Online. Sep 13 01:21:02.564634 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:21:02.576022 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 13 01:21:02.595162 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 13 01:21:03.348111 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:03.359881 (kubelet)[1921]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:21:03.791319 kubelet[1921]: E0913 01:21:03.791238 1921 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:21:03.792230 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:21:03.792310 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:21:04.518468 systemd-resolved[1731]: Clock change detected. Flushing caches. Sep 13 01:21:04.518479 systemd-timesyncd[1777]: Contacted time server 142.202.190.19:123 (0.flatcar.pool.ntp.org). Sep 13 01:21:04.518508 systemd-timesyncd[1777]: Initial clock synchronization to Sat 2025-09-13 01:21:04.518407 UTC. Sep 13 01:21:05.116915 kernel: mlx5_core 0000:01:00.0: lag map: port 1:1 port 2:2 Sep 13 01:21:05.117052 kernel: mlx5_core 0000:01:00.0: shared_fdb:0 mode:queue_affinity Sep 13 01:21:05.451392 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 13 01:21:05.462064 systemd[1]: Started sshd@0-139.178.94.15:22-139.178.89.65:44446.service - OpenSSH per-connection server daemon (139.178.89.65:44446). Sep 13 01:21:05.507750 coreos-metadata[1856]: Sep 13 01:21:05.507 INFO Fetch successful Sep 13 01:21:05.518072 sshd[1941]: Accepted publickey for core from 139.178.89.65 port 44446 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:05.518946 sshd[1941]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:05.522219 coreos-metadata[1782]: Sep 13 01:21:05.522 INFO Fetch successful Sep 13 01:21:05.524752 systemd-logind[1809]: New session 1 of user core. Sep 13 01:21:05.525695 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 13 01:21:05.542028 unknown[1856]: wrote ssh authorized keys file for user: core Sep 13 01:21:05.545483 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 13 01:21:05.559184 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 13 01:21:05.572428 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 13 01:21:05.580478 update-ssh-keys[1945]: Updated "/home/core/.ssh/authorized_keys" Sep 13 01:21:05.581070 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Sep 13 01:21:05.598469 systemd[1]: Finished sshkeys.service. Sep 13 01:21:05.605635 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Sep 13 01:21:05.627217 systemd[1]: Starting packet-phone-home.service - Report Success to Packet... Sep 13 01:21:05.628021 (systemd)[1949]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 13 01:21:05.710690 systemd[1949]: Queued start job for default target default.target. Sep 13 01:21:05.720803 systemd[1949]: Created slice app.slice - User Application Slice. Sep 13 01:21:05.720818 systemd[1949]: Reached target paths.target - Paths. Sep 13 01:21:05.720827 systemd[1949]: Reached target timers.target - Timers. Sep 13 01:21:05.721503 systemd[1949]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 13 01:21:05.727210 systemd[1949]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 13 01:21:05.727255 systemd[1949]: Reached target sockets.target - Sockets. Sep 13 01:21:05.727270 systemd[1949]: Reached target basic.target - Basic System. Sep 13 01:21:05.727298 systemd[1949]: Reached target default.target - Main User Target. Sep 13 01:21:05.727323 systemd[1949]: Startup finished in 87ms. Sep 13 01:21:05.727363 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 13 01:21:05.745383 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 13 01:21:05.805227 systemd[1]: Started sshd@1-139.178.94.15:22-139.178.89.65:44452.service - OpenSSH per-connection server daemon (139.178.89.65:44452). Sep 13 01:21:05.844398 sshd[1966]: Accepted publickey for core from 139.178.89.65 port 44452 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:05.845053 sshd[1966]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:05.847253 systemd-logind[1809]: New session 2 of user core. Sep 13 01:21:05.862383 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 13 01:21:05.918121 sshd[1966]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:05.929632 systemd[1]: sshd@1-139.178.94.15:22-139.178.89.65:44452.service: Deactivated successfully. Sep 13 01:21:05.930390 systemd[1]: session-2.scope: Deactivated successfully. Sep 13 01:21:05.930923 systemd-logind[1809]: Session 2 logged out. Waiting for processes to exit. Sep 13 01:21:05.931603 systemd[1]: Started sshd@2-139.178.94.15:22-139.178.89.65:44460.service - OpenSSH per-connection server daemon (139.178.89.65:44460). Sep 13 01:21:05.942876 systemd-logind[1809]: Removed session 2. Sep 13 01:21:05.970446 sshd[1973]: Accepted publickey for core from 139.178.89.65 port 44460 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:05.971095 sshd[1973]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:05.973479 systemd-logind[1809]: New session 3 of user core. Sep 13 01:21:05.984425 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 13 01:21:06.040123 sshd[1973]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:06.041331 systemd[1]: sshd@2-139.178.94.15:22-139.178.89.65:44460.service: Deactivated successfully. Sep 13 01:21:06.042102 systemd[1]: session-3.scope: Deactivated successfully. Sep 13 01:21:06.042685 systemd-logind[1809]: Session 3 logged out. Waiting for processes to exit. Sep 13 01:21:06.043169 systemd-logind[1809]: Removed session 3. Sep 13 01:21:06.068689 systemd[1]: Finished packet-phone-home.service - Report Success to Packet. Sep 13 01:21:06.079747 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 13 01:21:06.089389 systemd[1]: Startup finished in 2.104s (kernel) + 25.160s (initrd) + 10.568s (userspace) = 37.833s. Sep 13 01:21:06.128557 login[1902]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Sep 13 01:21:06.128766 login[1903]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Sep 13 01:21:06.131024 systemd-logind[1809]: New session 5 of user core. Sep 13 01:21:06.131657 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 13 01:21:06.132762 systemd-logind[1809]: New session 4 of user core. Sep 13 01:21:06.133284 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 13 01:21:14.315024 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 13 01:21:14.330535 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:21:14.572893 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:14.575038 (kubelet)[2014]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:21:14.605595 kubelet[2014]: E0913 01:21:14.605516 2014 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:21:14.607493 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:21:14.607571 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:21:16.062486 systemd[1]: Started sshd@3-139.178.94.15:22-139.178.89.65:34916.service - OpenSSH per-connection server daemon (139.178.89.65:34916). Sep 13 01:21:16.092538 sshd[2032]: Accepted publickey for core from 139.178.89.65 port 34916 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:16.093208 sshd[2032]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:16.095659 systemd-logind[1809]: New session 6 of user core. Sep 13 01:21:16.108705 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 13 01:21:16.168876 sshd[2032]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:16.187945 systemd[1]: sshd@3-139.178.94.15:22-139.178.89.65:34916.service: Deactivated successfully. Sep 13 01:21:16.188814 systemd[1]: session-6.scope: Deactivated successfully. Sep 13 01:21:16.189588 systemd-logind[1809]: Session 6 logged out. Waiting for processes to exit. Sep 13 01:21:16.190316 systemd[1]: Started sshd@4-139.178.94.15:22-139.178.89.65:34928.service - OpenSSH per-connection server daemon (139.178.89.65:34928). Sep 13 01:21:16.190998 systemd-logind[1809]: Removed session 6. Sep 13 01:21:16.243285 sshd[2039]: Accepted publickey for core from 139.178.89.65 port 34928 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:16.244645 sshd[2039]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:16.248866 systemd-logind[1809]: New session 7 of user core. Sep 13 01:21:16.265593 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 13 01:21:16.319340 sshd[2039]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:16.333880 systemd[1]: sshd@4-139.178.94.15:22-139.178.89.65:34928.service: Deactivated successfully. Sep 13 01:21:16.334690 systemd[1]: session-7.scope: Deactivated successfully. Sep 13 01:21:16.335475 systemd-logind[1809]: Session 7 logged out. Waiting for processes to exit. Sep 13 01:21:16.336207 systemd[1]: Started sshd@5-139.178.94.15:22-139.178.89.65:34942.service - OpenSSH per-connection server daemon (139.178.89.65:34942). Sep 13 01:21:16.336791 systemd-logind[1809]: Removed session 7. Sep 13 01:21:16.367874 sshd[2046]: Accepted publickey for core from 139.178.89.65 port 34942 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:16.368542 sshd[2046]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:16.371071 systemd-logind[1809]: New session 8 of user core. Sep 13 01:21:16.385518 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 13 01:21:16.438428 sshd[2046]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:16.466572 systemd[1]: sshd@5-139.178.94.15:22-139.178.89.65:34942.service: Deactivated successfully. Sep 13 01:21:16.470266 systemd[1]: session-8.scope: Deactivated successfully. Sep 13 01:21:16.473555 systemd-logind[1809]: Session 8 logged out. Waiting for processes to exit. Sep 13 01:21:16.488072 systemd[1]: Started sshd@6-139.178.94.15:22-139.178.89.65:34948.service - OpenSSH per-connection server daemon (139.178.89.65:34948). Sep 13 01:21:16.490521 systemd-logind[1809]: Removed session 8. Sep 13 01:21:16.557118 sshd[2053]: Accepted publickey for core from 139.178.89.65 port 34948 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:16.558142 sshd[2053]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:16.561697 systemd-logind[1809]: New session 9 of user core. Sep 13 01:21:16.572485 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 13 01:21:16.637442 sudo[2056]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 13 01:21:16.637595 sudo[2056]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:21:16.659116 sudo[2056]: pam_unix(sudo:session): session closed for user root Sep 13 01:21:16.660126 sshd[2053]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:16.671325 systemd[1]: sshd@6-139.178.94.15:22-139.178.89.65:34948.service: Deactivated successfully. Sep 13 01:21:16.672435 systemd[1]: session-9.scope: Deactivated successfully. Sep 13 01:21:16.673499 systemd-logind[1809]: Session 9 logged out. Waiting for processes to exit. Sep 13 01:21:16.674449 systemd[1]: Started sshd@7-139.178.94.15:22-139.178.89.65:34962.service - OpenSSH per-connection server daemon (139.178.89.65:34962). Sep 13 01:21:16.675205 systemd-logind[1809]: Removed session 9. Sep 13 01:21:16.734262 sshd[2061]: Accepted publickey for core from 139.178.89.65 port 34962 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:16.735663 sshd[2061]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:16.740288 systemd-logind[1809]: New session 10 of user core. Sep 13 01:21:16.754507 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 13 01:21:16.813282 sudo[2065]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 13 01:21:16.813434 sudo[2065]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:21:16.815439 sudo[2065]: pam_unix(sudo:session): session closed for user root Sep 13 01:21:16.818100 sudo[2064]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 13 01:21:16.818261 sudo[2064]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:21:16.836583 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 13 01:21:16.837651 auditctl[2068]: No rules Sep 13 01:21:16.837888 systemd[1]: audit-rules.service: Deactivated successfully. Sep 13 01:21:16.838016 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 13 01:21:16.839601 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 13 01:21:16.856054 augenrules[2086]: No rules Sep 13 01:21:16.856430 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 13 01:21:16.857010 sudo[2064]: pam_unix(sudo:session): session closed for user root Sep 13 01:21:16.857892 sshd[2061]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:16.874324 systemd[1]: sshd@7-139.178.94.15:22-139.178.89.65:34962.service: Deactivated successfully. Sep 13 01:21:16.875368 systemd[1]: session-10.scope: Deactivated successfully. Sep 13 01:21:16.876303 systemd-logind[1809]: Session 10 logged out. Waiting for processes to exit. Sep 13 01:21:16.877215 systemd[1]: Started sshd@8-139.178.94.15:22-139.178.89.65:34974.service - OpenSSH per-connection server daemon (139.178.89.65:34974). Sep 13 01:21:16.877969 systemd-logind[1809]: Removed session 10. Sep 13 01:21:16.934069 sshd[2094]: Accepted publickey for core from 139.178.89.65 port 34974 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:21:16.935289 sshd[2094]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:21:16.939340 systemd-logind[1809]: New session 11 of user core. Sep 13 01:21:16.950493 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 13 01:21:17.003348 sudo[2097]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 13 01:21:17.003504 sudo[2097]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:21:17.286696 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 13 01:21:17.286768 (dockerd)[2122]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 13 01:21:17.567961 dockerd[2122]: time="2025-09-13T01:21:17.567935674Z" level=info msg="Starting up" Sep 13 01:21:17.642570 dockerd[2122]: time="2025-09-13T01:21:17.642520371Z" level=info msg="Loading containers: start." Sep 13 01:21:17.715341 kernel: Initializing XFRM netlink socket Sep 13 01:21:17.800603 systemd-networkd[1608]: docker0: Link UP Sep 13 01:21:17.819220 dockerd[2122]: time="2025-09-13T01:21:17.819180533Z" level=info msg="Loading containers: done." Sep 13 01:21:17.828565 dockerd[2122]: time="2025-09-13T01:21:17.828520649Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 13 01:21:17.828538 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1196158224-merged.mount: Deactivated successfully. Sep 13 01:21:17.828692 dockerd[2122]: time="2025-09-13T01:21:17.828572955Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Sep 13 01:21:17.828692 dockerd[2122]: time="2025-09-13T01:21:17.828629803Z" level=info msg="Daemon has completed initialization" Sep 13 01:21:17.844166 dockerd[2122]: time="2025-09-13T01:21:17.844127599Z" level=info msg="API listen on /run/docker.sock" Sep 13 01:21:17.844254 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 13 01:21:18.657224 containerd[1819]: time="2025-09-13T01:21:18.657199955Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Sep 13 01:21:19.192789 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4177291744.mount: Deactivated successfully. Sep 13 01:21:19.956059 containerd[1819]: time="2025-09-13T01:21:19.956006465Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:19.956300 containerd[1819]: time="2025-09-13T01:21:19.956165072Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Sep 13 01:21:19.956697 containerd[1819]: time="2025-09-13T01:21:19.956657755Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:19.958586 containerd[1819]: time="2025-09-13T01:21:19.958542235Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:19.959109 containerd[1819]: time="2025-09-13T01:21:19.959082472Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 1.301860033s" Sep 13 01:21:19.959109 containerd[1819]: time="2025-09-13T01:21:19.959099416Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Sep 13 01:21:19.959421 containerd[1819]: time="2025-09-13T01:21:19.959407927Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Sep 13 01:21:20.897859 containerd[1819]: time="2025-09-13T01:21:20.897807633Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:20.898079 containerd[1819]: time="2025-09-13T01:21:20.898031967Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Sep 13 01:21:20.898420 containerd[1819]: time="2025-09-13T01:21:20.898380774Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:20.900392 containerd[1819]: time="2025-09-13T01:21:20.900350179Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:20.900894 containerd[1819]: time="2025-09-13T01:21:20.900852709Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 941.428322ms" Sep 13 01:21:20.900894 containerd[1819]: time="2025-09-13T01:21:20.900869627Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Sep 13 01:21:20.901197 containerd[1819]: time="2025-09-13T01:21:20.901157602Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Sep 13 01:21:21.776282 containerd[1819]: time="2025-09-13T01:21:21.776223906Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:21.776500 containerd[1819]: time="2025-09-13T01:21:21.776476101Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Sep 13 01:21:21.776861 containerd[1819]: time="2025-09-13T01:21:21.776821583Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:21.778485 containerd[1819]: time="2025-09-13T01:21:21.778441756Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:21.779174 containerd[1819]: time="2025-09-13T01:21:21.779131926Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 877.954844ms" Sep 13 01:21:21.779174 containerd[1819]: time="2025-09-13T01:21:21.779148907Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Sep 13 01:21:21.779439 containerd[1819]: time="2025-09-13T01:21:21.779396885Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Sep 13 01:21:22.751162 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4046994773.mount: Deactivated successfully. Sep 13 01:21:22.940017 containerd[1819]: time="2025-09-13T01:21:22.939967686Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:22.940254 containerd[1819]: time="2025-09-13T01:21:22.940180382Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Sep 13 01:21:22.940581 containerd[1819]: time="2025-09-13T01:21:22.940537844Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:22.941463 containerd[1819]: time="2025-09-13T01:21:22.941448091Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:22.942156 containerd[1819]: time="2025-09-13T01:21:22.942142134Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 1.16272507s" Sep 13 01:21:22.942183 containerd[1819]: time="2025-09-13T01:21:22.942160394Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Sep 13 01:21:22.942463 containerd[1819]: time="2025-09-13T01:21:22.942424324Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 13 01:21:23.384364 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount901081362.mount: Deactivated successfully. Sep 13 01:21:23.901381 containerd[1819]: time="2025-09-13T01:21:23.901323991Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:23.901583 containerd[1819]: time="2025-09-13T01:21:23.901535587Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Sep 13 01:21:23.902001 containerd[1819]: time="2025-09-13T01:21:23.901961442Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:23.903663 containerd[1819]: time="2025-09-13T01:21:23.903647248Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:23.904325 containerd[1819]: time="2025-09-13T01:21:23.904291164Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 961.850571ms" Sep 13 01:21:23.904325 containerd[1819]: time="2025-09-13T01:21:23.904307745Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 13 01:21:23.904591 containerd[1819]: time="2025-09-13T01:21:23.904580871Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 13 01:21:24.340763 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1777282440.mount: Deactivated successfully. Sep 13 01:21:24.342015 containerd[1819]: time="2025-09-13T01:21:24.341969290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:24.342179 containerd[1819]: time="2025-09-13T01:21:24.342136898Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 13 01:21:24.342613 containerd[1819]: time="2025-09-13T01:21:24.342577638Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:24.343672 containerd[1819]: time="2025-09-13T01:21:24.343628920Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:24.344157 containerd[1819]: time="2025-09-13T01:21:24.344121632Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 439.526757ms" Sep 13 01:21:24.344157 containerd[1819]: time="2025-09-13T01:21:24.344152318Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 13 01:21:24.344548 containerd[1819]: time="2025-09-13T01:21:24.344500078Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Sep 13 01:21:24.815167 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 13 01:21:24.831538 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:21:25.024045 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1323980046.mount: Deactivated successfully. Sep 13 01:21:25.106007 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:25.108511 (kubelet)[2441]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:21:25.149203 kubelet[2441]: E0913 01:21:25.149179 2441 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:21:25.150452 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:21:25.150549 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:21:26.250656 containerd[1819]: time="2025-09-13T01:21:26.250596383Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:26.250864 containerd[1819]: time="2025-09-13T01:21:26.250785423Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Sep 13 01:21:26.251361 containerd[1819]: time="2025-09-13T01:21:26.251310405Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:26.253389 containerd[1819]: time="2025-09-13T01:21:26.253348641Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:26.254000 containerd[1819]: time="2025-09-13T01:21:26.253956968Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 1.909441102s" Sep 13 01:21:26.254000 containerd[1819]: time="2025-09-13T01:21:26.253974204Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Sep 13 01:21:28.164445 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:28.178655 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:21:28.191744 systemd[1]: Reloading requested from client PID 2549 ('systemctl') (unit session-11.scope)... Sep 13 01:21:28.191752 systemd[1]: Reloading... Sep 13 01:21:28.233257 zram_generator::config[2588]: No configuration found. Sep 13 01:21:28.301882 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:21:28.362505 systemd[1]: Reloading finished in 170 ms. Sep 13 01:21:28.403397 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 13 01:21:28.403590 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 13 01:21:28.404033 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:28.416991 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:21:28.690123 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:28.694810 (kubelet)[2653]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 13 01:21:28.719124 kubelet[2653]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:21:28.719124 kubelet[2653]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 13 01:21:28.719124 kubelet[2653]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:21:28.719390 kubelet[2653]: I0913 01:21:28.719152 2653 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 01:21:28.911677 kubelet[2653]: I0913 01:21:28.911631 2653 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 13 01:21:28.911677 kubelet[2653]: I0913 01:21:28.911644 2653 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 01:21:28.912541 kubelet[2653]: I0913 01:21:28.912528 2653 server.go:954] "Client rotation is on, will bootstrap in background" Sep 13 01:21:28.932586 kubelet[2653]: E0913 01:21:28.932545 2653 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://139.178.94.15:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:28.933199 kubelet[2653]: I0913 01:21:28.933163 2653 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 01:21:28.938158 kubelet[2653]: E0913 01:21:28.938128 2653 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 01:21:28.938158 kubelet[2653]: I0913 01:21:28.938158 2653 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 01:21:28.946823 kubelet[2653]: I0913 01:21:28.946783 2653 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 01:21:28.947841 kubelet[2653]: I0913 01:21:28.947820 2653 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 01:21:28.947966 kubelet[2653]: I0913 01:21:28.947843 2653 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.5-n-7cd54b3887","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 13 01:21:28.948048 kubelet[2653]: I0913 01:21:28.947972 2653 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 01:21:28.948048 kubelet[2653]: I0913 01:21:28.947982 2653 container_manager_linux.go:304] "Creating device plugin manager" Sep 13 01:21:28.948114 kubelet[2653]: I0913 01:21:28.948062 2653 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:21:28.951413 kubelet[2653]: I0913 01:21:28.951402 2653 kubelet.go:446] "Attempting to sync node with API server" Sep 13 01:21:28.951498 kubelet[2653]: I0913 01:21:28.951418 2653 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 01:21:28.951498 kubelet[2653]: I0913 01:21:28.951432 2653 kubelet.go:352] "Adding apiserver pod source" Sep 13 01:21:28.951498 kubelet[2653]: I0913 01:21:28.951441 2653 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 01:21:28.954130 kubelet[2653]: I0913 01:21:28.954103 2653 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 13 01:21:28.954410 kubelet[2653]: I0913 01:21:28.954404 2653 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 01:21:28.955145 kubelet[2653]: W0913 01:21:28.955119 2653 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 13 01:21:28.955244 kubelet[2653]: W0913 01:21:28.955206 2653 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.94.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.5-n-7cd54b3887&limit=500&resourceVersion=0": dial tcp 139.178.94.15:6443: connect: connection refused Sep 13 01:21:28.955271 kubelet[2653]: E0913 01:21:28.955260 2653 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://139.178.94.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.5-n-7cd54b3887&limit=500&resourceVersion=0\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:28.956168 kubelet[2653]: W0913 01:21:28.956149 2653 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.94.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 139.178.94.15:6443: connect: connection refused Sep 13 01:21:28.956190 kubelet[2653]: E0913 01:21:28.956174 2653 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://139.178.94.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:28.956983 kubelet[2653]: I0913 01:21:28.956976 2653 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 13 01:21:28.957019 kubelet[2653]: I0913 01:21:28.956994 2653 server.go:1287] "Started kubelet" Sep 13 01:21:28.957899 kubelet[2653]: I0913 01:21:28.957884 2653 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 01:21:28.957946 kubelet[2653]: I0913 01:21:28.957938 2653 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 01:21:28.957968 kubelet[2653]: I0913 01:21:28.957944 2653 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 01:21:28.958044 kubelet[2653]: I0913 01:21:28.958035 2653 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 13 01:21:28.958079 kubelet[2653]: E0913 01:21:28.958062 2653 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.5-n-7cd54b3887\" not found" Sep 13 01:21:28.958116 kubelet[2653]: I0913 01:21:28.958106 2653 reconciler.go:26] "Reconciler: start to sync state" Sep 13 01:21:28.958116 kubelet[2653]: I0913 01:21:28.958106 2653 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 13 01:21:28.958166 kubelet[2653]: E0913 01:21:28.958143 2653 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.94.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-n-7cd54b3887?timeout=10s\": dial tcp 139.178.94.15:6443: connect: connection refused" interval="200ms" Sep 13 01:21:28.958235 kubelet[2653]: W0913 01:21:28.958217 2653 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.94.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.94.15:6443: connect: connection refused Sep 13 01:21:28.958276 kubelet[2653]: E0913 01:21:28.958239 2653 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://139.178.94.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:28.958276 kubelet[2653]: I0913 01:21:28.958256 2653 factory.go:221] Registration of the systemd container factory successfully Sep 13 01:21:28.958341 kubelet[2653]: I0913 01:21:28.958306 2653 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 01:21:28.979603 kubelet[2653]: I0913 01:21:28.979550 2653 factory.go:221] Registration of the containerd container factory successfully Sep 13 01:21:28.979855 kubelet[2653]: I0913 01:21:28.979688 2653 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 01:21:28.980004 kubelet[2653]: I0913 01:21:28.979993 2653 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 01:21:28.980048 kubelet[2653]: E0913 01:21:28.980030 2653 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 01:21:28.980350 kubelet[2653]: I0913 01:21:28.980341 2653 server.go:479] "Adding debug handlers to kubelet server" Sep 13 01:21:28.982471 kubelet[2653]: E0913 01:21:28.981330 2653 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.94.15:6443/api/v1/namespaces/default/events\": dial tcp 139.178.94.15:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.5-n-7cd54b3887.1864b2edca935512 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.5-n-7cd54b3887,UID:ci-4081.3.5-n-7cd54b3887,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.5-n-7cd54b3887,},FirstTimestamp:2025-09-13 01:21:28.956982546 +0000 UTC m=+0.259935917,LastTimestamp:2025-09-13 01:21:28.956982546 +0000 UTC m=+0.259935917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.5-n-7cd54b3887,}" Sep 13 01:21:28.987195 kubelet[2653]: I0913 01:21:28.987179 2653 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 01:21:28.987718 kubelet[2653]: I0913 01:21:28.987710 2653 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 01:21:28.987751 kubelet[2653]: I0913 01:21:28.987721 2653 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 13 01:21:28.987751 kubelet[2653]: I0913 01:21:28.987731 2653 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 13 01:21:28.987751 kubelet[2653]: I0913 01:21:28.987736 2653 kubelet.go:2382] "Starting kubelet main sync loop" Sep 13 01:21:28.987799 kubelet[2653]: E0913 01:21:28.987759 2653 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 01:21:28.987996 kubelet[2653]: W0913 01:21:28.987982 2653 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.94.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.94.15:6443: connect: connection refused Sep 13 01:21:28.988024 kubelet[2653]: E0913 01:21:28.988003 2653 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.94.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:29.053302 kubelet[2653]: I0913 01:21:29.053273 2653 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 13 01:21:29.053302 kubelet[2653]: I0913 01:21:29.053298 2653 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 13 01:21:29.053471 kubelet[2653]: I0913 01:21:29.053317 2653 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:21:29.054651 kubelet[2653]: I0913 01:21:29.054610 2653 policy_none.go:49] "None policy: Start" Sep 13 01:21:29.054651 kubelet[2653]: I0913 01:21:29.054631 2653 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 13 01:21:29.054651 kubelet[2653]: I0913 01:21:29.054644 2653 state_mem.go:35] "Initializing new in-memory state store" Sep 13 01:21:29.057572 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 13 01:21:29.058898 kubelet[2653]: E0913 01:21:29.058888 2653 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.5-n-7cd54b3887\" not found" Sep 13 01:21:29.072853 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 13 01:21:29.074428 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 13 01:21:29.087989 kubelet[2653]: E0913 01:21:29.087962 2653 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 13 01:21:29.092848 kubelet[2653]: I0913 01:21:29.092811 2653 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 01:21:29.092955 kubelet[2653]: I0913 01:21:29.092899 2653 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 01:21:29.092955 kubelet[2653]: I0913 01:21:29.092907 2653 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 01:21:29.093030 kubelet[2653]: I0913 01:21:29.093014 2653 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 01:21:29.093346 kubelet[2653]: E0913 01:21:29.093312 2653 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 13 01:21:29.093346 kubelet[2653]: E0913 01:21:29.093335 2653 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.5-n-7cd54b3887\" not found" Sep 13 01:21:29.160371 kubelet[2653]: E0913 01:21:29.160158 2653 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.94.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-n-7cd54b3887?timeout=10s\": dial tcp 139.178.94.15:6443: connect: connection refused" interval="400ms" Sep 13 01:21:29.197867 kubelet[2653]: I0913 01:21:29.197636 2653 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.198518 kubelet[2653]: E0913 01:21:29.198410 2653 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.94.15:6443/api/v1/nodes\": dial tcp 139.178.94.15:6443: connect: connection refused" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.258818 kubelet[2653]: E0913 01:21:29.258690 2653 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.94.15:6443/api/v1/namespaces/default/events\": dial tcp 139.178.94.15:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.5-n-7cd54b3887.1864b2edca935512 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.5-n-7cd54b3887,UID:ci-4081.3.5-n-7cd54b3887,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.5-n-7cd54b3887,},FirstTimestamp:2025-09-13 01:21:28.956982546 +0000 UTC m=+0.259935917,LastTimestamp:2025-09-13 01:21:28.956982546 +0000 UTC m=+0.259935917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.5-n-7cd54b3887,}" Sep 13 01:21:29.310473 systemd[1]: Created slice kubepods-burstable-pod141448e41dcf9818c12ed425a530f9ba.slice - libcontainer container kubepods-burstable-pod141448e41dcf9818c12ed425a530f9ba.slice. Sep 13 01:21:29.330126 kubelet[2653]: E0913 01:21:29.330044 2653 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-n-7cd54b3887\" not found" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.335429 systemd[1]: Created slice kubepods-burstable-podc99f55c7c6a3edc7c044ffd2747db592.slice - libcontainer container kubepods-burstable-podc99f55c7c6a3edc7c044ffd2747db592.slice. Sep 13 01:21:29.348330 kubelet[2653]: E0913 01:21:29.348228 2653 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-n-7cd54b3887\" not found" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.355433 systemd[1]: Created slice kubepods-burstable-pod1774ee678989d9cfa2a29146fc68657c.slice - libcontainer container kubepods-burstable-pod1774ee678989d9cfa2a29146fc68657c.slice. Sep 13 01:21:29.359782 kubelet[2653]: E0913 01:21:29.359703 2653 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-n-7cd54b3887\" not found" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360325 kubelet[2653]: I0913 01:21:29.360220 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360513 kubelet[2653]: I0913 01:21:29.360305 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360513 kubelet[2653]: I0913 01:21:29.360393 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c99f55c7c6a3edc7c044ffd2747db592-kubeconfig\") pod \"kube-scheduler-ci-4081.3.5-n-7cd54b3887\" (UID: \"c99f55c7c6a3edc7c044ffd2747db592\") " pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360513 kubelet[2653]: I0913 01:21:29.360441 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1774ee678989d9cfa2a29146fc68657c-k8s-certs\") pod \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" (UID: \"1774ee678989d9cfa2a29146fc68657c\") " pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360513 kubelet[2653]: I0913 01:21:29.360490 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1774ee678989d9cfa2a29146fc68657c-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" (UID: \"1774ee678989d9cfa2a29146fc68657c\") " pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360860 kubelet[2653]: I0913 01:21:29.360536 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-ca-certs\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360860 kubelet[2653]: I0913 01:21:29.360587 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360860 kubelet[2653]: I0913 01:21:29.360634 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.360860 kubelet[2653]: I0913 01:21:29.360693 2653 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1774ee678989d9cfa2a29146fc68657c-ca-certs\") pod \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" (UID: \"1774ee678989d9cfa2a29146fc68657c\") " pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.406557 kubelet[2653]: I0913 01:21:29.406457 2653 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.407440 kubelet[2653]: E0913 01:21:29.407323 2653 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.94.15:6443/api/v1/nodes\": dial tcp 139.178.94.15:6443: connect: connection refused" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.561400 kubelet[2653]: E0913 01:21:29.561145 2653 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.94.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-n-7cd54b3887?timeout=10s\": dial tcp 139.178.94.15:6443: connect: connection refused" interval="800ms" Sep 13 01:21:29.633153 containerd[1819]: time="2025-09-13T01:21:29.633025299Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.5-n-7cd54b3887,Uid:141448e41dcf9818c12ed425a530f9ba,Namespace:kube-system,Attempt:0,}" Sep 13 01:21:29.650864 containerd[1819]: time="2025-09-13T01:21:29.650836992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.5-n-7cd54b3887,Uid:c99f55c7c6a3edc7c044ffd2747db592,Namespace:kube-system,Attempt:0,}" Sep 13 01:21:29.660883 containerd[1819]: time="2025-09-13T01:21:29.660832400Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.5-n-7cd54b3887,Uid:1774ee678989d9cfa2a29146fc68657c,Namespace:kube-system,Attempt:0,}" Sep 13 01:21:29.812202 kubelet[2653]: I0913 01:21:29.812004 2653 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.812948 kubelet[2653]: E0913 01:21:29.812704 2653 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.94.15:6443/api/v1/nodes\": dial tcp 139.178.94.15:6443: connect: connection refused" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:29.897385 kubelet[2653]: W0913 01:21:29.897286 2653 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.94.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 139.178.94.15:6443: connect: connection refused Sep 13 01:21:29.897385 kubelet[2653]: E0913 01:21:29.897368 2653 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://139.178.94.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:29.980173 kubelet[2653]: W0913 01:21:29.980096 2653 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.94.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.94.15:6443: connect: connection refused Sep 13 01:21:29.980173 kubelet[2653]: E0913 01:21:29.980149 2653 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.94.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:29.999567 kubelet[2653]: W0913 01:21:29.999385 2653 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.94.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.94.15:6443: connect: connection refused Sep 13 01:21:29.999567 kubelet[2653]: E0913 01:21:29.999540 2653 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://139.178.94.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.94.15:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:21:30.149087 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount433020529.mount: Deactivated successfully. Sep 13 01:21:30.150820 containerd[1819]: time="2025-09-13T01:21:30.150773852Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:21:30.150970 containerd[1819]: time="2025-09-13T01:21:30.150915865Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Sep 13 01:21:30.151425 containerd[1819]: time="2025-09-13T01:21:30.151388165Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:21:30.151746 containerd[1819]: time="2025-09-13T01:21:30.151702352Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 13 01:21:30.151846 containerd[1819]: time="2025-09-13T01:21:30.151796451Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:21:30.152243 containerd[1819]: time="2025-09-13T01:21:30.152193238Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 13 01:21:30.152243 containerd[1819]: time="2025-09-13T01:21:30.152193899Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:21:30.154386 containerd[1819]: time="2025-09-13T01:21:30.154310728Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:21:30.155295 containerd[1819]: time="2025-09-13T01:21:30.155205806Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 494.345782ms" Sep 13 01:21:30.155682 containerd[1819]: time="2025-09-13T01:21:30.155638296Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 522.459325ms" Sep 13 01:21:30.157282 containerd[1819]: time="2025-09-13T01:21:30.157215323Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 506.332774ms" Sep 13 01:21:30.274707 containerd[1819]: time="2025-09-13T01:21:30.274642484Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:21:30.274707 containerd[1819]: time="2025-09-13T01:21:30.274687595Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:21:30.274707 containerd[1819]: time="2025-09-13T01:21:30.274703752Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:30.274842 containerd[1819]: time="2025-09-13T01:21:30.274821044Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:30.274945 containerd[1819]: time="2025-09-13T01:21:30.274704907Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:21:30.274999 containerd[1819]: time="2025-09-13T01:21:30.274947451Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:21:30.274999 containerd[1819]: time="2025-09-13T01:21:30.274957590Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:30.275042 containerd[1819]: time="2025-09-13T01:21:30.274999240Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:30.275252 containerd[1819]: time="2025-09-13T01:21:30.275220662Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:21:30.275279 containerd[1819]: time="2025-09-13T01:21:30.275252548Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:21:30.275279 containerd[1819]: time="2025-09-13T01:21:30.275262258Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:30.275320 containerd[1819]: time="2025-09-13T01:21:30.275298510Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:30.296547 systemd[1]: Started cri-containerd-92771c4a8256c1cc94cd4b3fb284555595c58a56fccc037b5a66075443bbc0c5.scope - libcontainer container 92771c4a8256c1cc94cd4b3fb284555595c58a56fccc037b5a66075443bbc0c5. Sep 13 01:21:30.297375 systemd[1]: Started cri-containerd-96683e70816621a47dae2e758f78136d5f4e50108e178ccd3c042a154cd72bdf.scope - libcontainer container 96683e70816621a47dae2e758f78136d5f4e50108e178ccd3c042a154cd72bdf. Sep 13 01:21:30.298176 systemd[1]: Started cri-containerd-b9bb2f28d8cc5d2258e5d21ec411415558cb1eba67d9db1de476468b814b44fa.scope - libcontainer container b9bb2f28d8cc5d2258e5d21ec411415558cb1eba67d9db1de476468b814b44fa. Sep 13 01:21:30.320918 containerd[1819]: time="2025-09-13T01:21:30.320890830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.5-n-7cd54b3887,Uid:141448e41dcf9818c12ed425a530f9ba,Namespace:kube-system,Attempt:0,} returns sandbox id \"92771c4a8256c1cc94cd4b3fb284555595c58a56fccc037b5a66075443bbc0c5\"" Sep 13 01:21:30.321014 containerd[1819]: time="2025-09-13T01:21:30.320985415Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.5-n-7cd54b3887,Uid:1774ee678989d9cfa2a29146fc68657c,Namespace:kube-system,Attempt:0,} returns sandbox id \"96683e70816621a47dae2e758f78136d5f4e50108e178ccd3c042a154cd72bdf\"" Sep 13 01:21:30.321460 containerd[1819]: time="2025-09-13T01:21:30.321444831Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.5-n-7cd54b3887,Uid:c99f55c7c6a3edc7c044ffd2747db592,Namespace:kube-system,Attempt:0,} returns sandbox id \"b9bb2f28d8cc5d2258e5d21ec411415558cb1eba67d9db1de476468b814b44fa\"" Sep 13 01:21:30.322510 containerd[1819]: time="2025-09-13T01:21:30.322497363Z" level=info msg="CreateContainer within sandbox \"92771c4a8256c1cc94cd4b3fb284555595c58a56fccc037b5a66075443bbc0c5\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 13 01:21:30.322552 containerd[1819]: time="2025-09-13T01:21:30.322498873Z" level=info msg="CreateContainer within sandbox \"b9bb2f28d8cc5d2258e5d21ec411415558cb1eba67d9db1de476468b814b44fa\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 13 01:21:30.322605 containerd[1819]: time="2025-09-13T01:21:30.322498871Z" level=info msg="CreateContainer within sandbox \"96683e70816621a47dae2e758f78136d5f4e50108e178ccd3c042a154cd72bdf\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 13 01:21:30.329702 containerd[1819]: time="2025-09-13T01:21:30.329685177Z" level=info msg="CreateContainer within sandbox \"92771c4a8256c1cc94cd4b3fb284555595c58a56fccc037b5a66075443bbc0c5\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"7cc2bc7179987ceb21e36fe476df42e6307581b660b7d36527f7f1acdb363b74\"" Sep 13 01:21:30.329988 containerd[1819]: time="2025-09-13T01:21:30.329976966Z" level=info msg="StartContainer for \"7cc2bc7179987ceb21e36fe476df42e6307581b660b7d36527f7f1acdb363b74\"" Sep 13 01:21:30.331355 containerd[1819]: time="2025-09-13T01:21:30.331337495Z" level=info msg="CreateContainer within sandbox \"96683e70816621a47dae2e758f78136d5f4e50108e178ccd3c042a154cd72bdf\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"267ff1070cad139088d23c85fa2aaf934c4e5703ef07d1e5e975f579a53bf59d\"" Sep 13 01:21:30.331501 containerd[1819]: time="2025-09-13T01:21:30.331488241Z" level=info msg="CreateContainer within sandbox \"b9bb2f28d8cc5d2258e5d21ec411415558cb1eba67d9db1de476468b814b44fa\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"be7469eb2c355e9fa748cf3dd0c6436237eb0ab98e843e91c67a52265a6fede8\"" Sep 13 01:21:30.331526 containerd[1819]: time="2025-09-13T01:21:30.331515572Z" level=info msg="StartContainer for \"267ff1070cad139088d23c85fa2aaf934c4e5703ef07d1e5e975f579a53bf59d\"" Sep 13 01:21:30.331717 containerd[1819]: time="2025-09-13T01:21:30.331706133Z" level=info msg="StartContainer for \"be7469eb2c355e9fa748cf3dd0c6436237eb0ab98e843e91c67a52265a6fede8\"" Sep 13 01:21:30.358394 systemd[1]: Started cri-containerd-267ff1070cad139088d23c85fa2aaf934c4e5703ef07d1e5e975f579a53bf59d.scope - libcontainer container 267ff1070cad139088d23c85fa2aaf934c4e5703ef07d1e5e975f579a53bf59d. Sep 13 01:21:30.358985 systemd[1]: Started cri-containerd-7cc2bc7179987ceb21e36fe476df42e6307581b660b7d36527f7f1acdb363b74.scope - libcontainer container 7cc2bc7179987ceb21e36fe476df42e6307581b660b7d36527f7f1acdb363b74. Sep 13 01:21:30.359547 systemd[1]: Started cri-containerd-be7469eb2c355e9fa748cf3dd0c6436237eb0ab98e843e91c67a52265a6fede8.scope - libcontainer container be7469eb2c355e9fa748cf3dd0c6436237eb0ab98e843e91c67a52265a6fede8. Sep 13 01:21:30.362251 kubelet[2653]: E0913 01:21:30.362226 2653 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.94.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-n-7cd54b3887?timeout=10s\": dial tcp 139.178.94.15:6443: connect: connection refused" interval="1.6s" Sep 13 01:21:30.382629 containerd[1819]: time="2025-09-13T01:21:30.382603069Z" level=info msg="StartContainer for \"267ff1070cad139088d23c85fa2aaf934c4e5703ef07d1e5e975f579a53bf59d\" returns successfully" Sep 13 01:21:30.382807 containerd[1819]: time="2025-09-13T01:21:30.382668376Z" level=info msg="StartContainer for \"be7469eb2c355e9fa748cf3dd0c6436237eb0ab98e843e91c67a52265a6fede8\" returns successfully" Sep 13 01:21:30.384070 containerd[1819]: time="2025-09-13T01:21:30.384051019Z" level=info msg="StartContainer for \"7cc2bc7179987ceb21e36fe476df42e6307581b660b7d36527f7f1acdb363b74\" returns successfully" Sep 13 01:21:30.614427 kubelet[2653]: I0913 01:21:30.614411 2653 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.935699 kubelet[2653]: I0913 01:21:30.935621 2653 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.953547 kubelet[2653]: I0913 01:21:30.953531 2653 apiserver.go:52] "Watching apiserver" Sep 13 01:21:30.958728 kubelet[2653]: I0913 01:21:30.958689 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.958728 kubelet[2653]: I0913 01:21:30.958726 2653 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 13 01:21:30.964193 kubelet[2653]: E0913 01:21:30.964179 2653 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.964226 kubelet[2653]: I0913 01:21:30.964194 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.964896 kubelet[2653]: E0913 01:21:30.964887 2653 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.964936 kubelet[2653]: I0913 01:21:30.964896 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.965530 kubelet[2653]: E0913 01:21:30.965520 2653 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.5-n-7cd54b3887\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.990696 kubelet[2653]: I0913 01:21:30.990682 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.991748 kubelet[2653]: E0913 01:21:30.991738 2653 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.5-n-7cd54b3887\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.992230 kubelet[2653]: I0913 01:21:30.992223 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.992742 kubelet[2653]: I0913 01:21:30.992736 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.993020 kubelet[2653]: E0913 01:21:30.993008 2653 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:30.993470 kubelet[2653]: E0913 01:21:30.993461 2653 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:31.994958 kubelet[2653]: I0913 01:21:31.994862 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:31.995760 kubelet[2653]: I0913 01:21:31.995100 2653 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:32.001311 kubelet[2653]: W0913 01:21:32.001239 2653 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:21:32.001789 kubelet[2653]: W0913 01:21:32.001741 2653 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:21:33.142252 systemd[1]: Reloading requested from client PID 2969 ('systemctl') (unit session-11.scope)... Sep 13 01:21:33.142260 systemd[1]: Reloading... Sep 13 01:21:33.180260 zram_generator::config[3008]: No configuration found. Sep 13 01:21:33.248155 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:21:33.316641 systemd[1]: Reloading finished in 174 ms. Sep 13 01:21:33.349751 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:21:33.360571 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 01:21:33.361053 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:33.379588 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:21:33.669382 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:21:33.672487 (kubelet)[3072]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 13 01:21:33.694759 kubelet[3072]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:21:33.694759 kubelet[3072]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 13 01:21:33.694759 kubelet[3072]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:21:33.694993 kubelet[3072]: I0913 01:21:33.694780 3072 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 01:21:33.699179 kubelet[3072]: I0913 01:21:33.699137 3072 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 13 01:21:33.699179 kubelet[3072]: I0913 01:21:33.699150 3072 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 01:21:33.699345 kubelet[3072]: I0913 01:21:33.699302 3072 server.go:954] "Client rotation is on, will bootstrap in background" Sep 13 01:21:33.700102 kubelet[3072]: I0913 01:21:33.700063 3072 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 13 01:21:33.701427 kubelet[3072]: I0913 01:21:33.701390 3072 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 01:21:33.703308 kubelet[3072]: E0913 01:21:33.703284 3072 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 01:21:33.703354 kubelet[3072]: I0913 01:21:33.703309 3072 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 01:21:33.711002 kubelet[3072]: I0913 01:21:33.710961 3072 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 01:21:33.711121 kubelet[3072]: I0913 01:21:33.711078 3072 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 01:21:33.711203 kubelet[3072]: I0913 01:21:33.711098 3072 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.5-n-7cd54b3887","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 13 01:21:33.711265 kubelet[3072]: I0913 01:21:33.711210 3072 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 01:21:33.711265 kubelet[3072]: I0913 01:21:33.711217 3072 container_manager_linux.go:304] "Creating device plugin manager" Sep 13 01:21:33.711265 kubelet[3072]: I0913 01:21:33.711253 3072 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:21:33.711403 kubelet[3072]: I0913 01:21:33.711365 3072 kubelet.go:446] "Attempting to sync node with API server" Sep 13 01:21:33.711403 kubelet[3072]: I0913 01:21:33.711377 3072 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 01:21:33.711403 kubelet[3072]: I0913 01:21:33.711388 3072 kubelet.go:352] "Adding apiserver pod source" Sep 13 01:21:33.711403 kubelet[3072]: I0913 01:21:33.711394 3072 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 01:21:33.711861 kubelet[3072]: I0913 01:21:33.711815 3072 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 13 01:21:33.712092 kubelet[3072]: I0913 01:21:33.712082 3072 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 01:21:33.712362 kubelet[3072]: I0913 01:21:33.712353 3072 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 13 01:21:33.712407 kubelet[3072]: I0913 01:21:33.712376 3072 server.go:1287] "Started kubelet" Sep 13 01:21:33.712506 kubelet[3072]: I0913 01:21:33.712471 3072 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 01:21:33.712575 kubelet[3072]: I0913 01:21:33.712528 3072 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 01:21:33.712710 kubelet[3072]: I0913 01:21:33.712700 3072 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 01:21:33.713558 kubelet[3072]: I0913 01:21:33.713550 3072 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 01:21:33.713706 kubelet[3072]: I0913 01:21:33.713684 3072 server.go:479] "Adding debug handlers to kubelet server" Sep 13 01:21:33.713837 kubelet[3072]: I0913 01:21:33.713745 3072 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 01:21:33.713837 kubelet[3072]: I0913 01:21:33.713773 3072 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 13 01:21:33.713923 kubelet[3072]: I0913 01:21:33.713759 3072 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 13 01:21:33.714217 kubelet[3072]: I0913 01:21:33.714206 3072 reconciler.go:26] "Reconciler: start to sync state" Sep 13 01:21:33.714350 kubelet[3072]: E0913 01:21:33.714333 3072 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.5-n-7cd54b3887\" not found" Sep 13 01:21:33.714554 kubelet[3072]: I0913 01:21:33.714543 3072 factory.go:221] Registration of the systemd container factory successfully Sep 13 01:21:33.715113 kubelet[3072]: I0913 01:21:33.715093 3072 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 01:21:33.715223 kubelet[3072]: E0913 01:21:33.715195 3072 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 01:21:33.715767 kubelet[3072]: I0913 01:21:33.715756 3072 factory.go:221] Registration of the containerd container factory successfully Sep 13 01:21:33.719595 kubelet[3072]: I0913 01:21:33.719571 3072 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 01:21:33.720263 kubelet[3072]: I0913 01:21:33.720253 3072 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 01:21:33.720302 kubelet[3072]: I0913 01:21:33.720271 3072 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 13 01:21:33.720302 kubelet[3072]: I0913 01:21:33.720283 3072 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 13 01:21:33.720302 kubelet[3072]: I0913 01:21:33.720289 3072 kubelet.go:2382] "Starting kubelet main sync loop" Sep 13 01:21:33.720365 kubelet[3072]: E0913 01:21:33.720318 3072 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 01:21:33.732109 kubelet[3072]: I0913 01:21:33.732091 3072 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 13 01:21:33.732109 kubelet[3072]: I0913 01:21:33.732105 3072 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 13 01:21:33.732216 kubelet[3072]: I0913 01:21:33.732118 3072 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:21:33.732240 kubelet[3072]: I0913 01:21:33.732219 3072 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 13 01:21:33.732240 kubelet[3072]: I0913 01:21:33.732230 3072 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 13 01:21:33.732290 kubelet[3072]: I0913 01:21:33.732242 3072 policy_none.go:49] "None policy: Start" Sep 13 01:21:33.732290 kubelet[3072]: I0913 01:21:33.732258 3072 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 13 01:21:33.732290 kubelet[3072]: I0913 01:21:33.732266 3072 state_mem.go:35] "Initializing new in-memory state store" Sep 13 01:21:33.732347 kubelet[3072]: I0913 01:21:33.732332 3072 state_mem.go:75] "Updated machine memory state" Sep 13 01:21:33.734442 kubelet[3072]: I0913 01:21:33.734430 3072 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 01:21:33.734536 kubelet[3072]: I0913 01:21:33.734528 3072 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 01:21:33.734572 kubelet[3072]: I0913 01:21:33.734536 3072 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 01:21:33.734655 kubelet[3072]: I0913 01:21:33.734631 3072 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 01:21:33.734996 kubelet[3072]: E0913 01:21:33.734983 3072 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 13 01:21:33.821927 kubelet[3072]: I0913 01:21:33.821859 3072 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:33.822320 kubelet[3072]: I0913 01:21:33.821965 3072 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:33.822320 kubelet[3072]: I0913 01:21:33.821871 3072 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:33.842180 kubelet[3072]: I0913 01:21:33.842107 3072 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:33.845451 kubelet[3072]: W0913 01:21:33.845391 3072 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:21:33.845691 kubelet[3072]: W0913 01:21:33.845502 3072 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:21:33.845691 kubelet[3072]: E0913 01:21:33.845559 3072 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.5-n-7cd54b3887\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:33.846034 kubelet[3072]: W0913 01:21:33.845906 3072 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:21:33.846034 kubelet[3072]: E0913 01:21:33.846015 3072 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:33.851709 kubelet[3072]: I0913 01:21:33.851620 3072 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:33.851907 kubelet[3072]: I0913 01:21:33.851782 3072 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016432 kubelet[3072]: I0913 01:21:34.016165 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1774ee678989d9cfa2a29146fc68657c-k8s-certs\") pod \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" (UID: \"1774ee678989d9cfa2a29146fc68657c\") " pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016432 kubelet[3072]: I0913 01:21:34.016305 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1774ee678989d9cfa2a29146fc68657c-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" (UID: \"1774ee678989d9cfa2a29146fc68657c\") " pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016432 kubelet[3072]: I0913 01:21:34.016388 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-ca-certs\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016876 kubelet[3072]: I0913 01:21:34.016454 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016876 kubelet[3072]: I0913 01:21:34.016506 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016876 kubelet[3072]: I0913 01:21:34.016559 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016876 kubelet[3072]: I0913 01:21:34.016618 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c99f55c7c6a3edc7c044ffd2747db592-kubeconfig\") pod \"kube-scheduler-ci-4081.3.5-n-7cd54b3887\" (UID: \"c99f55c7c6a3edc7c044ffd2747db592\") " pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.016876 kubelet[3072]: I0913 01:21:34.016668 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1774ee678989d9cfa2a29146fc68657c-ca-certs\") pod \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" (UID: \"1774ee678989d9cfa2a29146fc68657c\") " pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.017416 kubelet[3072]: I0913 01:21:34.016719 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/141448e41dcf9818c12ed425a530f9ba-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.5-n-7cd54b3887\" (UID: \"141448e41dcf9818c12ed425a530f9ba\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.712471 kubelet[3072]: I0913 01:21:34.712453 3072 apiserver.go:52] "Watching apiserver" Sep 13 01:21:34.714790 kubelet[3072]: I0913 01:21:34.714779 3072 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 13 01:21:34.726154 kubelet[3072]: I0913 01:21:34.726143 3072 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.728851 kubelet[3072]: W0913 01:21:34.728818 3072 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:21:34.728851 kubelet[3072]: E0913 01:21:34.728841 3072 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-n-7cd54b3887\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" Sep 13 01:21:34.740313 kubelet[3072]: I0913 01:21:34.740258 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.5-n-7cd54b3887" podStartSLOduration=3.740242081 podStartE2EDuration="3.740242081s" podCreationTimestamp="2025-09-13 01:21:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:21:34.736297774 +0000 UTC m=+1.061549219" watchObservedRunningTime="2025-09-13 01:21:34.740242081 +0000 UTC m=+1.065493525" Sep 13 01:21:34.744384 kubelet[3072]: I0913 01:21:34.744334 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.5-n-7cd54b3887" podStartSLOduration=1.7443255579999999 podStartE2EDuration="1.744325558s" podCreationTimestamp="2025-09-13 01:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:21:34.744324869 +0000 UTC m=+1.069576316" watchObservedRunningTime="2025-09-13 01:21:34.744325558 +0000 UTC m=+1.069577001" Sep 13 01:21:34.744384 kubelet[3072]: I0913 01:21:34.744375 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.5-n-7cd54b3887" podStartSLOduration=3.744371479 podStartE2EDuration="3.744371479s" podCreationTimestamp="2025-09-13 01:21:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:21:34.740337926 +0000 UTC m=+1.065589370" watchObservedRunningTime="2025-09-13 01:21:34.744371479 +0000 UTC m=+1.069622921" Sep 13 01:21:38.295166 kubelet[3072]: I0913 01:21:38.295086 3072 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 13 01:21:38.296206 containerd[1819]: time="2025-09-13T01:21:38.295780680Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 13 01:21:38.297146 kubelet[3072]: I0913 01:21:38.296337 3072 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 13 01:21:38.749397 kubelet[3072]: I0913 01:21:38.749272 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/654a320f-2b4b-4fc8-aa42-b49e83baf165-kube-proxy\") pod \"kube-proxy-9t78l\" (UID: \"654a320f-2b4b-4fc8-aa42-b49e83baf165\") " pod="kube-system/kube-proxy-9t78l" Sep 13 01:21:38.749815 kubelet[3072]: I0913 01:21:38.749415 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/654a320f-2b4b-4fc8-aa42-b49e83baf165-xtables-lock\") pod \"kube-proxy-9t78l\" (UID: \"654a320f-2b4b-4fc8-aa42-b49e83baf165\") " pod="kube-system/kube-proxy-9t78l" Sep 13 01:21:38.749815 kubelet[3072]: I0913 01:21:38.749525 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfsx4\" (UniqueName: \"kubernetes.io/projected/654a320f-2b4b-4fc8-aa42-b49e83baf165-kube-api-access-kfsx4\") pod \"kube-proxy-9t78l\" (UID: \"654a320f-2b4b-4fc8-aa42-b49e83baf165\") " pod="kube-system/kube-proxy-9t78l" Sep 13 01:21:38.749815 kubelet[3072]: I0913 01:21:38.749633 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/654a320f-2b4b-4fc8-aa42-b49e83baf165-lib-modules\") pod \"kube-proxy-9t78l\" (UID: \"654a320f-2b4b-4fc8-aa42-b49e83baf165\") " pod="kube-system/kube-proxy-9t78l" Sep 13 01:21:38.750012 systemd[1]: Created slice kubepods-besteffort-pod654a320f_2b4b_4fc8_aa42_b49e83baf165.slice - libcontainer container kubepods-besteffort-pod654a320f_2b4b_4fc8_aa42_b49e83baf165.slice. Sep 13 01:21:38.864301 kubelet[3072]: E0913 01:21:38.864209 3072 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 13 01:21:38.864301 kubelet[3072]: E0913 01:21:38.864300 3072 projected.go:194] Error preparing data for projected volume kube-api-access-kfsx4 for pod kube-system/kube-proxy-9t78l: configmap "kube-root-ca.crt" not found Sep 13 01:21:38.864681 kubelet[3072]: E0913 01:21:38.864438 3072 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/654a320f-2b4b-4fc8-aa42-b49e83baf165-kube-api-access-kfsx4 podName:654a320f-2b4b-4fc8-aa42-b49e83baf165 nodeName:}" failed. No retries permitted until 2025-09-13 01:21:39.364390329 +0000 UTC m=+5.689641844 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-kfsx4" (UniqueName: "kubernetes.io/projected/654a320f-2b4b-4fc8-aa42-b49e83baf165-kube-api-access-kfsx4") pod "kube-proxy-9t78l" (UID: "654a320f-2b4b-4fc8-aa42-b49e83baf165") : configmap "kube-root-ca.crt" not found Sep 13 01:21:39.444243 systemd[1]: Created slice kubepods-besteffort-pod3e39b53f_19c1_471d_93a6_4000a69de720.slice - libcontainer container kubepods-besteffort-pod3e39b53f_19c1_471d_93a6_4000a69de720.slice. Sep 13 01:21:39.455215 kubelet[3072]: I0913 01:21:39.455093 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3e39b53f-19c1-471d-93a6-4000a69de720-var-lib-calico\") pod \"tigera-operator-755d956888-wbpcg\" (UID: \"3e39b53f-19c1-471d-93a6-4000a69de720\") " pod="tigera-operator/tigera-operator-755d956888-wbpcg" Sep 13 01:21:39.455215 kubelet[3072]: I0913 01:21:39.455199 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tn9v\" (UniqueName: \"kubernetes.io/projected/3e39b53f-19c1-471d-93a6-4000a69de720-kube-api-access-6tn9v\") pod \"tigera-operator-755d956888-wbpcg\" (UID: \"3e39b53f-19c1-471d-93a6-4000a69de720\") " pod="tigera-operator/tigera-operator-755d956888-wbpcg" Sep 13 01:21:39.671646 containerd[1819]: time="2025-09-13T01:21:39.671557630Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-9t78l,Uid:654a320f-2b4b-4fc8-aa42-b49e83baf165,Namespace:kube-system,Attempt:0,}" Sep 13 01:21:39.707100 containerd[1819]: time="2025-09-13T01:21:39.706775063Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:21:39.707100 containerd[1819]: time="2025-09-13T01:21:39.706979376Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:21:39.707100 containerd[1819]: time="2025-09-13T01:21:39.706987355Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:39.707100 containerd[1819]: time="2025-09-13T01:21:39.707041802Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:39.721416 systemd[1]: Started cri-containerd-3641699dab5209fa33bcebe7ee3562961697f3fc32b7495f8998aedb8c3796b9.scope - libcontainer container 3641699dab5209fa33bcebe7ee3562961697f3fc32b7495f8998aedb8c3796b9. Sep 13 01:21:39.731114 containerd[1819]: time="2025-09-13T01:21:39.731065399Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-9t78l,Uid:654a320f-2b4b-4fc8-aa42-b49e83baf165,Namespace:kube-system,Attempt:0,} returns sandbox id \"3641699dab5209fa33bcebe7ee3562961697f3fc32b7495f8998aedb8c3796b9\"" Sep 13 01:21:39.732344 containerd[1819]: time="2025-09-13T01:21:39.732328572Z" level=info msg="CreateContainer within sandbox \"3641699dab5209fa33bcebe7ee3562961697f3fc32b7495f8998aedb8c3796b9\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 13 01:21:39.751303 containerd[1819]: time="2025-09-13T01:21:39.751206688Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-wbpcg,Uid:3e39b53f-19c1-471d-93a6-4000a69de720,Namespace:tigera-operator,Attempt:0,}" Sep 13 01:21:39.775089 containerd[1819]: time="2025-09-13T01:21:39.775038344Z" level=info msg="CreateContainer within sandbox \"3641699dab5209fa33bcebe7ee3562961697f3fc32b7495f8998aedb8c3796b9\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8a488437d277cf184bfd02abdd5b8acd5386b4b0cb18e4b76710bd8c35bca073\"" Sep 13 01:21:39.775408 containerd[1819]: time="2025-09-13T01:21:39.775390722Z" level=info msg="StartContainer for \"8a488437d277cf184bfd02abdd5b8acd5386b4b0cb18e4b76710bd8c35bca073\"" Sep 13 01:21:39.779728 containerd[1819]: time="2025-09-13T01:21:39.779651522Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:21:39.779728 containerd[1819]: time="2025-09-13T01:21:39.779693211Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:21:39.779921 containerd[1819]: time="2025-09-13T01:21:39.779905225Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:39.779961 containerd[1819]: time="2025-09-13T01:21:39.779951278Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:39.794462 systemd[1]: Started cri-containerd-0f15e53b15b5592730f8e9fa01af826b27d789770ad35c6fdbe1b7148b110cd9.scope - libcontainer container 0f15e53b15b5592730f8e9fa01af826b27d789770ad35c6fdbe1b7148b110cd9. Sep 13 01:21:39.795097 systemd[1]: Started cri-containerd-8a488437d277cf184bfd02abdd5b8acd5386b4b0cb18e4b76710bd8c35bca073.scope - libcontainer container 8a488437d277cf184bfd02abdd5b8acd5386b4b0cb18e4b76710bd8c35bca073. Sep 13 01:21:39.870334 containerd[1819]: time="2025-09-13T01:21:39.870288090Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-wbpcg,Uid:3e39b53f-19c1-471d-93a6-4000a69de720,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"0f15e53b15b5592730f8e9fa01af826b27d789770ad35c6fdbe1b7148b110cd9\"" Sep 13 01:21:39.870334 containerd[1819]: time="2025-09-13T01:21:39.870329960Z" level=info msg="StartContainer for \"8a488437d277cf184bfd02abdd5b8acd5386b4b0cb18e4b76710bd8c35bca073\" returns successfully" Sep 13 01:21:39.871782 containerd[1819]: time="2025-09-13T01:21:39.871761604Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 13 01:21:41.856515 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount439346069.mount: Deactivated successfully. Sep 13 01:21:42.330580 containerd[1819]: time="2025-09-13T01:21:42.330555641Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:42.330829 containerd[1819]: time="2025-09-13T01:21:42.330802649Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=25062609" Sep 13 01:21:42.331223 containerd[1819]: time="2025-09-13T01:21:42.331209513Z" level=info msg="ImageCreate event name:\"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:42.332313 containerd[1819]: time="2025-09-13T01:21:42.332271584Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:42.333113 containerd[1819]: time="2025-09-13T01:21:42.333077742Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"25058604\" in 2.461291795s" Sep 13 01:21:42.333113 containerd[1819]: time="2025-09-13T01:21:42.333092541Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\"" Sep 13 01:21:42.334034 containerd[1819]: time="2025-09-13T01:21:42.333994787Z" level=info msg="CreateContainer within sandbox \"0f15e53b15b5592730f8e9fa01af826b27d789770ad35c6fdbe1b7148b110cd9\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 13 01:21:42.338243 containerd[1819]: time="2025-09-13T01:21:42.338193544Z" level=info msg="CreateContainer within sandbox \"0f15e53b15b5592730f8e9fa01af826b27d789770ad35c6fdbe1b7148b110cd9\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"4465a9dc8d007d6a57d1fed34a0229211f14fda7a6e7b1a3697926b27fd6f4ec\"" Sep 13 01:21:42.338506 containerd[1819]: time="2025-09-13T01:21:42.338463213Z" level=info msg="StartContainer for \"4465a9dc8d007d6a57d1fed34a0229211f14fda7a6e7b1a3697926b27fd6f4ec\"" Sep 13 01:21:42.353570 systemd[1]: Started cri-containerd-4465a9dc8d007d6a57d1fed34a0229211f14fda7a6e7b1a3697926b27fd6f4ec.scope - libcontainer container 4465a9dc8d007d6a57d1fed34a0229211f14fda7a6e7b1a3697926b27fd6f4ec. Sep 13 01:21:42.364467 containerd[1819]: time="2025-09-13T01:21:42.364416737Z" level=info msg="StartContainer for \"4465a9dc8d007d6a57d1fed34a0229211f14fda7a6e7b1a3697926b27fd6f4ec\" returns successfully" Sep 13 01:21:42.767639 kubelet[3072]: I0913 01:21:42.767360 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-9t78l" podStartSLOduration=4.767312437 podStartE2EDuration="4.767312437s" podCreationTimestamp="2025-09-13 01:21:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:21:40.7561173 +0000 UTC m=+7.081368817" watchObservedRunningTime="2025-09-13 01:21:42.767312437 +0000 UTC m=+9.092563978" Sep 13 01:21:42.768588 kubelet[3072]: I0913 01:21:42.767746 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-wbpcg" podStartSLOduration=1.305642523 podStartE2EDuration="3.767724387s" podCreationTimestamp="2025-09-13 01:21:39 +0000 UTC" firstStartedPulling="2025-09-13 01:21:39.871375109 +0000 UTC m=+6.196626567" lastFinishedPulling="2025-09-13 01:21:42.33345699 +0000 UTC m=+8.658708431" observedRunningTime="2025-09-13 01:21:42.767612298 +0000 UTC m=+9.092863835" watchObservedRunningTime="2025-09-13 01:21:42.767724387 +0000 UTC m=+9.092975885" Sep 13 01:21:46.858016 sudo[2097]: pam_unix(sudo:session): session closed for user root Sep 13 01:21:46.858818 sshd[2094]: pam_unix(sshd:session): session closed for user core Sep 13 01:21:46.860697 systemd[1]: sshd@8-139.178.94.15:22-139.178.89.65:34974.service: Deactivated successfully. Sep 13 01:21:46.861902 systemd[1]: session-11.scope: Deactivated successfully. Sep 13 01:21:46.862046 systemd[1]: session-11.scope: Consumed 3.510s CPU time, 168.0M memory peak, 0B memory swap peak. Sep 13 01:21:46.864484 systemd-logind[1809]: Session 11 logged out. Waiting for processes to exit. Sep 13 01:21:46.869035 systemd-logind[1809]: Removed session 11. Sep 13 01:21:47.257176 update_engine[1814]: I20250913 01:21:47.257081 1814 update_attempter.cc:509] Updating boot flags... Sep 13 01:21:47.285259 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 36 scanned by (udev-worker) (3601) Sep 13 01:21:47.312258 kernel: BTRFS warning: duplicate device /dev/sdb3 devid 1 generation 36 scanned by (udev-worker) (3604) Sep 13 01:21:49.057365 systemd[1]: Created slice kubepods-besteffort-podbb35245b_998b_4b1f_adaa_fcab876e0c78.slice - libcontainer container kubepods-besteffort-podbb35245b_998b_4b1f_adaa_fcab876e0c78.slice. Sep 13 01:21:49.124361 kubelet[3072]: I0913 01:21:49.124205 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7sfs\" (UniqueName: \"kubernetes.io/projected/bb35245b-998b-4b1f-adaa-fcab876e0c78-kube-api-access-f7sfs\") pod \"calico-typha-78d45fb74c-hw9vw\" (UID: \"bb35245b-998b-4b1f-adaa-fcab876e0c78\") " pod="calico-system/calico-typha-78d45fb74c-hw9vw" Sep 13 01:21:49.125214 kubelet[3072]: I0913 01:21:49.124402 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb35245b-998b-4b1f-adaa-fcab876e0c78-tigera-ca-bundle\") pod \"calico-typha-78d45fb74c-hw9vw\" (UID: \"bb35245b-998b-4b1f-adaa-fcab876e0c78\") " pod="calico-system/calico-typha-78d45fb74c-hw9vw" Sep 13 01:21:49.125214 kubelet[3072]: I0913 01:21:49.124461 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/bb35245b-998b-4b1f-adaa-fcab876e0c78-typha-certs\") pod \"calico-typha-78d45fb74c-hw9vw\" (UID: \"bb35245b-998b-4b1f-adaa-fcab876e0c78\") " pod="calico-system/calico-typha-78d45fb74c-hw9vw" Sep 13 01:21:49.361554 containerd[1819]: time="2025-09-13T01:21:49.361430134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-78d45fb74c-hw9vw,Uid:bb35245b-998b-4b1f-adaa-fcab876e0c78,Namespace:calico-system,Attempt:0,}" Sep 13 01:21:49.371818 containerd[1819]: time="2025-09-13T01:21:49.371745694Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:21:49.371985 containerd[1819]: time="2025-09-13T01:21:49.371972006Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:21:49.372014 containerd[1819]: time="2025-09-13T01:21:49.371981828Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:49.372033 containerd[1819]: time="2025-09-13T01:21:49.372023415Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:49.387435 systemd[1]: Started cri-containerd-172a1cf72ee7245dbe458f8c9cbc126edaa00cb3e3c326fed7954225d34c7171.scope - libcontainer container 172a1cf72ee7245dbe458f8c9cbc126edaa00cb3e3c326fed7954225d34c7171. Sep 13 01:21:49.412428 containerd[1819]: time="2025-09-13T01:21:49.412380766Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-78d45fb74c-hw9vw,Uid:bb35245b-998b-4b1f-adaa-fcab876e0c78,Namespace:calico-system,Attempt:0,} returns sandbox id \"172a1cf72ee7245dbe458f8c9cbc126edaa00cb3e3c326fed7954225d34c7171\"" Sep 13 01:21:49.413085 containerd[1819]: time="2025-09-13T01:21:49.413072300Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 13 01:21:49.431422 systemd[1]: Created slice kubepods-besteffort-pod1f9609f0_671e_4d0b_a7d3_a258fd393631.slice - libcontainer container kubepods-besteffort-pod1f9609f0_671e_4d0b_a7d3_a258fd393631.slice. Sep 13 01:21:49.527855 kubelet[3072]: I0913 01:21:49.527745 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-cni-bin-dir\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.527855 kubelet[3072]: I0913 01:21:49.527849 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f9609f0-671e-4d0b-a7d3-a258fd393631-tigera-ca-bundle\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528199 kubelet[3072]: I0913 01:21:49.527917 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-flexvol-driver-host\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528199 kubelet[3072]: I0913 01:21:49.528090 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/1f9609f0-671e-4d0b-a7d3-a258fd393631-node-certs\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528455 kubelet[3072]: I0913 01:21:49.528218 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-var-run-calico\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528455 kubelet[3072]: I0913 01:21:49.528323 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-policysync\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528455 kubelet[3072]: I0913 01:21:49.528393 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-cni-net-dir\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528718 kubelet[3072]: I0913 01:21:49.528456 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-lib-modules\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528718 kubelet[3072]: I0913 01:21:49.528527 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-xtables-lock\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528718 kubelet[3072]: I0913 01:21:49.528598 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-var-lib-calico\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.528718 kubelet[3072]: I0913 01:21:49.528665 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmt6p\" (UniqueName: \"kubernetes.io/projected/1f9609f0-671e-4d0b-a7d3-a258fd393631-kube-api-access-wmt6p\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.529068 kubelet[3072]: I0913 01:21:49.528733 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/1f9609f0-671e-4d0b-a7d3-a258fd393631-cni-log-dir\") pod \"calico-node-lxfxq\" (UID: \"1f9609f0-671e-4d0b-a7d3-a258fd393631\") " pod="calico-system/calico-node-lxfxq" Sep 13 01:21:49.632766 kubelet[3072]: E0913 01:21:49.632575 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.632766 kubelet[3072]: W0913 01:21:49.632616 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.632766 kubelet[3072]: E0913 01:21:49.632682 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.636702 kubelet[3072]: E0913 01:21:49.636654 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.636702 kubelet[3072]: W0913 01:21:49.636690 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.637275 kubelet[3072]: E0913 01:21:49.636726 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.648751 kubelet[3072]: E0913 01:21:49.648655 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.648751 kubelet[3072]: W0913 01:21:49.648696 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.648751 kubelet[3072]: E0913 01:21:49.648730 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.734393 containerd[1819]: time="2025-09-13T01:21:49.734300635Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-lxfxq,Uid:1f9609f0-671e-4d0b-a7d3-a258fd393631,Namespace:calico-system,Attempt:0,}" Sep 13 01:21:49.740947 kubelet[3072]: E0913 01:21:49.740913 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-jqdjm" podUID="f4d72f52-2d93-4ed1-9fa8-3baa96684daf" Sep 13 01:21:49.758469 containerd[1819]: time="2025-09-13T01:21:49.758389368Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:21:49.758469 containerd[1819]: time="2025-09-13T01:21:49.758419472Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:21:49.758641 containerd[1819]: time="2025-09-13T01:21:49.758432107Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:49.758684 containerd[1819]: time="2025-09-13T01:21:49.758670133Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:21:49.774429 systemd[1]: Started cri-containerd-36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0.scope - libcontainer container 36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0. Sep 13 01:21:49.786378 containerd[1819]: time="2025-09-13T01:21:49.786354022Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-lxfxq,Uid:1f9609f0-671e-4d0b-a7d3-a258fd393631,Namespace:calico-system,Attempt:0,} returns sandbox id \"36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0\"" Sep 13 01:21:49.816253 kubelet[3072]: E0913 01:21:49.816233 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.816325 kubelet[3072]: W0913 01:21:49.816255 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.816325 kubelet[3072]: E0913 01:21:49.816269 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.816386 kubelet[3072]: E0913 01:21:49.816379 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.816386 kubelet[3072]: W0913 01:21:49.816386 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.816432 kubelet[3072]: E0913 01:21:49.816391 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.816513 kubelet[3072]: E0913 01:21:49.816507 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.816513 kubelet[3072]: W0913 01:21:49.816512 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.816556 kubelet[3072]: E0913 01:21:49.816518 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.816639 kubelet[3072]: E0913 01:21:49.816633 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.816639 kubelet[3072]: W0913 01:21:49.816638 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.816683 kubelet[3072]: E0913 01:21:49.816644 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.816772 kubelet[3072]: E0913 01:21:49.816766 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.816772 kubelet[3072]: W0913 01:21:49.816771 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.816810 kubelet[3072]: E0913 01:21:49.816776 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.816891 kubelet[3072]: E0913 01:21:49.816887 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.816911 kubelet[3072]: W0913 01:21:49.816891 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.816911 kubelet[3072]: E0913 01:21:49.816896 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817005 kubelet[3072]: E0913 01:21:49.817001 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817005 kubelet[3072]: W0913 01:21:49.817005 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817043 kubelet[3072]: E0913 01:21:49.817010 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817095 kubelet[3072]: E0913 01:21:49.817090 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817114 kubelet[3072]: W0913 01:21:49.817095 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817114 kubelet[3072]: E0913 01:21:49.817099 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817181 kubelet[3072]: E0913 01:21:49.817176 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817199 kubelet[3072]: W0913 01:21:49.817181 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817199 kubelet[3072]: E0913 01:21:49.817185 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817278 kubelet[3072]: E0913 01:21:49.817273 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817278 kubelet[3072]: W0913 01:21:49.817278 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817319 kubelet[3072]: E0913 01:21:49.817282 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817356 kubelet[3072]: E0913 01:21:49.817352 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817376 kubelet[3072]: W0913 01:21:49.817356 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817376 kubelet[3072]: E0913 01:21:49.817360 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817436 kubelet[3072]: E0913 01:21:49.817431 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817453 kubelet[3072]: W0913 01:21:49.817436 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817453 kubelet[3072]: E0913 01:21:49.817441 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817531 kubelet[3072]: E0913 01:21:49.817526 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817531 kubelet[3072]: W0913 01:21:49.817531 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817567 kubelet[3072]: E0913 01:21:49.817535 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817604 kubelet[3072]: E0913 01:21:49.817599 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817604 kubelet[3072]: W0913 01:21:49.817604 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817642 kubelet[3072]: E0913 01:21:49.817608 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817677 kubelet[3072]: E0913 01:21:49.817672 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817697 kubelet[3072]: W0913 01:21:49.817676 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817697 kubelet[3072]: E0913 01:21:49.817681 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817749 kubelet[3072]: E0913 01:21:49.817744 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817770 kubelet[3072]: W0913 01:21:49.817749 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817770 kubelet[3072]: E0913 01:21:49.817753 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817826 kubelet[3072]: E0913 01:21:49.817822 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817847 kubelet[3072]: W0913 01:21:49.817826 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817847 kubelet[3072]: E0913 01:21:49.817830 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817903 kubelet[3072]: E0913 01:21:49.817898 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817924 kubelet[3072]: W0913 01:21:49.817903 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.817924 kubelet[3072]: E0913 01:21:49.817907 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.817975 kubelet[3072]: E0913 01:21:49.817970 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.817975 kubelet[3072]: W0913 01:21:49.817974 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.818008 kubelet[3072]: E0913 01:21:49.817979 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.818055 kubelet[3072]: E0913 01:21:49.818051 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.818055 kubelet[3072]: W0913 01:21:49.818055 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.818091 kubelet[3072]: E0913 01:21:49.818059 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.818163 kubelet[3072]: E0913 01:21:49.818157 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.818189 kubelet[3072]: W0913 01:21:49.818163 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.818189 kubelet[3072]: E0913 01:21:49.818168 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.818260 kubelet[3072]: E0913 01:21:49.818252 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.818260 kubelet[3072]: W0913 01:21:49.818257 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.818330 kubelet[3072]: E0913 01:21:49.818262 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.818350 kubelet[3072]: E0913 01:21:49.818342 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.818350 kubelet[3072]: W0913 01:21:49.818346 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.818387 kubelet[3072]: E0913 01:21:49.818350 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.818435 kubelet[3072]: E0913 01:21:49.818430 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.818435 kubelet[3072]: W0913 01:21:49.818434 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.818473 kubelet[3072]: E0913 01:21:49.818439 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.818530 kubelet[3072]: E0913 01:21:49.818525 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.818549 kubelet[3072]: W0913 01:21:49.818531 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.818549 kubelet[3072]: E0913 01:21:49.818537 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832048 kubelet[3072]: E0913 01:21:49.832008 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832048 kubelet[3072]: W0913 01:21:49.832021 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832048 kubelet[3072]: E0913 01:21:49.832032 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832048 kubelet[3072]: I0913 01:21:49.832048 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2mgg\" (UniqueName: \"kubernetes.io/projected/f4d72f52-2d93-4ed1-9fa8-3baa96684daf-kube-api-access-g2mgg\") pod \"csi-node-driver-jqdjm\" (UID: \"f4d72f52-2d93-4ed1-9fa8-3baa96684daf\") " pod="calico-system/csi-node-driver-jqdjm" Sep 13 01:21:49.832187 kubelet[3072]: E0913 01:21:49.832173 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832187 kubelet[3072]: W0913 01:21:49.832179 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832187 kubelet[3072]: E0913 01:21:49.832186 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832234 kubelet[3072]: I0913 01:21:49.832195 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f4d72f52-2d93-4ed1-9fa8-3baa96684daf-registration-dir\") pod \"csi-node-driver-jqdjm\" (UID: \"f4d72f52-2d93-4ed1-9fa8-3baa96684daf\") " pod="calico-system/csi-node-driver-jqdjm" Sep 13 01:21:49.832368 kubelet[3072]: E0913 01:21:49.832329 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832368 kubelet[3072]: W0913 01:21:49.832335 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832368 kubelet[3072]: E0913 01:21:49.832341 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832368 kubelet[3072]: I0913 01:21:49.832350 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f4d72f52-2d93-4ed1-9fa8-3baa96684daf-kubelet-dir\") pod \"csi-node-driver-jqdjm\" (UID: \"f4d72f52-2d93-4ed1-9fa8-3baa96684daf\") " pod="calico-system/csi-node-driver-jqdjm" Sep 13 01:21:49.832530 kubelet[3072]: E0913 01:21:49.832493 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832530 kubelet[3072]: W0913 01:21:49.832499 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832530 kubelet[3072]: E0913 01:21:49.832505 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832530 kubelet[3072]: I0913 01:21:49.832514 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/f4d72f52-2d93-4ed1-9fa8-3baa96684daf-varrun\") pod \"csi-node-driver-jqdjm\" (UID: \"f4d72f52-2d93-4ed1-9fa8-3baa96684daf\") " pod="calico-system/csi-node-driver-jqdjm" Sep 13 01:21:49.832622 kubelet[3072]: E0913 01:21:49.832616 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832641 kubelet[3072]: W0913 01:21:49.832622 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832641 kubelet[3072]: E0913 01:21:49.832628 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832641 kubelet[3072]: I0913 01:21:49.832636 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f4d72f52-2d93-4ed1-9fa8-3baa96684daf-socket-dir\") pod \"csi-node-driver-jqdjm\" (UID: \"f4d72f52-2d93-4ed1-9fa8-3baa96684daf\") " pod="calico-system/csi-node-driver-jqdjm" Sep 13 01:21:49.832749 kubelet[3072]: E0913 01:21:49.832743 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832766 kubelet[3072]: W0913 01:21:49.832750 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832766 kubelet[3072]: E0913 01:21:49.832758 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832843 kubelet[3072]: E0913 01:21:49.832838 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832861 kubelet[3072]: W0913 01:21:49.832843 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832861 kubelet[3072]: E0913 01:21:49.832849 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.832970 kubelet[3072]: E0913 01:21:49.832941 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.832970 kubelet[3072]: W0913 01:21:49.832946 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.832970 kubelet[3072]: E0913 01:21:49.832952 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.833028 kubelet[3072]: E0913 01:21:49.833023 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.833049 kubelet[3072]: W0913 01:21:49.833028 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.833049 kubelet[3072]: E0913 01:21:49.833034 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.833119 kubelet[3072]: E0913 01:21:49.833115 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.833140 kubelet[3072]: W0913 01:21:49.833119 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.833140 kubelet[3072]: E0913 01:21:49.833125 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.833198 kubelet[3072]: E0913 01:21:49.833192 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.833198 kubelet[3072]: W0913 01:21:49.833196 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.833230 kubelet[3072]: E0913 01:21:49.833202 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.833286 kubelet[3072]: E0913 01:21:49.833282 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.833286 kubelet[3072]: W0913 01:21:49.833286 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.833322 kubelet[3072]: E0913 01:21:49.833292 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.833406 kubelet[3072]: E0913 01:21:49.833401 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.833425 kubelet[3072]: W0913 01:21:49.833406 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.833425 kubelet[3072]: E0913 01:21:49.833411 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.833534 kubelet[3072]: E0913 01:21:49.833529 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.833553 kubelet[3072]: W0913 01:21:49.833533 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.833553 kubelet[3072]: E0913 01:21:49.833537 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.833610 kubelet[3072]: E0913 01:21:49.833605 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.833629 kubelet[3072]: W0913 01:21:49.833610 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.833629 kubelet[3072]: E0913 01:21:49.833614 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.934146 kubelet[3072]: E0913 01:21:49.933940 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.934146 kubelet[3072]: W0913 01:21:49.933986 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.934146 kubelet[3072]: E0913 01:21:49.934027 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.934734 kubelet[3072]: E0913 01:21:49.934657 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.934734 kubelet[3072]: W0913 01:21:49.934686 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.934734 kubelet[3072]: E0913 01:21:49.934725 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.935315 kubelet[3072]: E0913 01:21:49.935266 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.935315 kubelet[3072]: W0913 01:21:49.935308 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.935617 kubelet[3072]: E0913 01:21:49.935347 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.936117 kubelet[3072]: E0913 01:21:49.936075 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.936281 kubelet[3072]: W0913 01:21:49.936122 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.936281 kubelet[3072]: E0913 01:21:49.936181 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.936654 kubelet[3072]: E0913 01:21:49.936645 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.936679 kubelet[3072]: W0913 01:21:49.936655 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.936679 kubelet[3072]: E0913 01:21:49.936672 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.936884 kubelet[3072]: E0913 01:21:49.936811 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.936884 kubelet[3072]: W0913 01:21:49.936820 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.936884 kubelet[3072]: E0913 01:21:49.936835 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.936952 kubelet[3072]: E0913 01:21:49.936948 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.936975 kubelet[3072]: W0913 01:21:49.936956 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.936975 kubelet[3072]: E0913 01:21:49.936969 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.937098 kubelet[3072]: E0913 01:21:49.937090 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.937123 kubelet[3072]: W0913 01:21:49.937099 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.937146 kubelet[3072]: E0913 01:21:49.937126 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.937266 kubelet[3072]: E0913 01:21:49.937258 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.937293 kubelet[3072]: W0913 01:21:49.937267 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.937293 kubelet[3072]: E0913 01:21:49.937281 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.937477 kubelet[3072]: E0913 01:21:49.937470 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.937506 kubelet[3072]: W0913 01:21:49.937479 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.937506 kubelet[3072]: E0913 01:21:49.937491 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.937676 kubelet[3072]: E0913 01:21:49.937642 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.937676 kubelet[3072]: W0913 01:21:49.937651 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.937676 kubelet[3072]: E0913 01:21:49.937663 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.937839 kubelet[3072]: E0913 01:21:49.937806 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.937839 kubelet[3072]: W0913 01:21:49.937814 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.937839 kubelet[3072]: E0913 01:21:49.937834 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.938042 kubelet[3072]: E0913 01:21:49.938009 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.938042 kubelet[3072]: W0913 01:21:49.938018 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.938042 kubelet[3072]: E0913 01:21:49.938039 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.938213 kubelet[3072]: E0913 01:21:49.938204 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.938213 kubelet[3072]: W0913 01:21:49.938213 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.938287 kubelet[3072]: E0913 01:21:49.938236 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.938413 kubelet[3072]: E0913 01:21:49.938403 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.938451 kubelet[3072]: W0913 01:21:49.938414 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.938451 kubelet[3072]: E0913 01:21:49.938425 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.938540 kubelet[3072]: E0913 01:21:49.938533 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.938540 kubelet[3072]: W0913 01:21:49.938540 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.938589 kubelet[3072]: E0913 01:21:49.938558 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.938658 kubelet[3072]: E0913 01:21:49.938652 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.938685 kubelet[3072]: W0913 01:21:49.938659 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.938685 kubelet[3072]: E0913 01:21:49.938672 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.938764 kubelet[3072]: E0913 01:21:49.938757 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.938791 kubelet[3072]: W0913 01:21:49.938763 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.938791 kubelet[3072]: E0913 01:21:49.938771 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.938922 kubelet[3072]: E0913 01:21:49.938913 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.938946 kubelet[3072]: W0913 01:21:49.938923 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.938946 kubelet[3072]: E0913 01:21:49.938933 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.939085 kubelet[3072]: E0913 01:21:49.939078 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.939114 kubelet[3072]: W0913 01:21:49.939085 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.939114 kubelet[3072]: E0913 01:21:49.939094 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.939236 kubelet[3072]: E0913 01:21:49.939229 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.939236 kubelet[3072]: W0913 01:21:49.939236 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.939302 kubelet[3072]: E0913 01:21:49.939250 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.939397 kubelet[3072]: E0913 01:21:49.939390 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.939424 kubelet[3072]: W0913 01:21:49.939397 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.939424 kubelet[3072]: E0913 01:21:49.939406 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.939580 kubelet[3072]: E0913 01:21:49.939573 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.939606 kubelet[3072]: W0913 01:21:49.939580 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.939606 kubelet[3072]: E0913 01:21:49.939588 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.939785 kubelet[3072]: E0913 01:21:49.939775 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.939821 kubelet[3072]: W0913 01:21:49.939787 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.939821 kubelet[3072]: E0913 01:21:49.939798 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.940018 kubelet[3072]: E0913 01:21:49.940007 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.940018 kubelet[3072]: W0913 01:21:49.940018 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.940083 kubelet[3072]: E0913 01:21:49.940028 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:49.946430 kubelet[3072]: E0913 01:21:49.946417 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:49.946430 kubelet[3072]: W0913 01:21:49.946429 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:49.946430 kubelet[3072]: E0913 01:21:49.946443 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:50.979382 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3872502972.mount: Deactivated successfully. Sep 13 01:21:51.638320 containerd[1819]: time="2025-09-13T01:21:51.638295276Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:51.638585 containerd[1819]: time="2025-09-13T01:21:51.638497968Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=35237389" Sep 13 01:21:51.638922 containerd[1819]: time="2025-09-13T01:21:51.638909055Z" level=info msg="ImageCreate event name:\"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:51.639848 containerd[1819]: time="2025-09-13T01:21:51.639832616Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:51.640265 containerd[1819]: time="2025-09-13T01:21:51.640252155Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"35237243\" in 2.227160087s" Sep 13 01:21:51.640305 containerd[1819]: time="2025-09-13T01:21:51.640267589Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\"" Sep 13 01:21:51.640741 containerd[1819]: time="2025-09-13T01:21:51.640730392Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 13 01:21:51.643535 containerd[1819]: time="2025-09-13T01:21:51.643487762Z" level=info msg="CreateContainer within sandbox \"172a1cf72ee7245dbe458f8c9cbc126edaa00cb3e3c326fed7954225d34c7171\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 13 01:21:51.647730 containerd[1819]: time="2025-09-13T01:21:51.647714771Z" level=info msg="CreateContainer within sandbox \"172a1cf72ee7245dbe458f8c9cbc126edaa00cb3e3c326fed7954225d34c7171\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"f5843fb3a9a22f056b4d70816ebf1e47b4f33f2094490d63441a8bfeeefd291a\"" Sep 13 01:21:51.647915 containerd[1819]: time="2025-09-13T01:21:51.647903619Z" level=info msg="StartContainer for \"f5843fb3a9a22f056b4d70816ebf1e47b4f33f2094490d63441a8bfeeefd291a\"" Sep 13 01:21:51.675700 systemd[1]: Started cri-containerd-f5843fb3a9a22f056b4d70816ebf1e47b4f33f2094490d63441a8bfeeefd291a.scope - libcontainer container f5843fb3a9a22f056b4d70816ebf1e47b4f33f2094490d63441a8bfeeefd291a. Sep 13 01:21:51.721356 kubelet[3072]: E0913 01:21:51.721229 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-jqdjm" podUID="f4d72f52-2d93-4ed1-9fa8-3baa96684daf" Sep 13 01:21:51.763813 containerd[1819]: time="2025-09-13T01:21:51.763777230Z" level=info msg="StartContainer for \"f5843fb3a9a22f056b4d70816ebf1e47b4f33f2094490d63441a8bfeeefd291a\" returns successfully" Sep 13 01:21:51.775452 kubelet[3072]: I0913 01:21:51.775348 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-78d45fb74c-hw9vw" podStartSLOduration=0.547618731 podStartE2EDuration="2.775327492s" podCreationTimestamp="2025-09-13 01:21:49 +0000 UTC" firstStartedPulling="2025-09-13 01:21:49.412946671 +0000 UTC m=+15.738198116" lastFinishedPulling="2025-09-13 01:21:51.640655434 +0000 UTC m=+17.965906877" observedRunningTime="2025-09-13 01:21:51.775257999 +0000 UTC m=+18.100509460" watchObservedRunningTime="2025-09-13 01:21:51.775327492 +0000 UTC m=+18.100578949" Sep 13 01:21:51.829707 kubelet[3072]: E0913 01:21:51.829675 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.829707 kubelet[3072]: W0913 01:21:51.829702 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.829882 kubelet[3072]: E0913 01:21:51.829725 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.830009 kubelet[3072]: E0913 01:21:51.829994 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.830009 kubelet[3072]: W0913 01:21:51.830008 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.830121 kubelet[3072]: E0913 01:21:51.830019 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.830237 kubelet[3072]: E0913 01:21:51.830226 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.830308 kubelet[3072]: W0913 01:21:51.830237 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.830355 kubelet[3072]: E0913 01:21:51.830319 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.830666 kubelet[3072]: E0913 01:21:51.830652 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.830666 kubelet[3072]: W0913 01:21:51.830664 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.830767 kubelet[3072]: E0913 01:21:51.830675 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.830938 kubelet[3072]: E0913 01:21:51.830923 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.830991 kubelet[3072]: W0913 01:21:51.830939 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.830991 kubelet[3072]: E0913 01:21:51.830956 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.831218 kubelet[3072]: E0913 01:21:51.831203 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.831291 kubelet[3072]: W0913 01:21:51.831220 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.831291 kubelet[3072]: E0913 01:21:51.831261 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.831519 kubelet[3072]: E0913 01:21:51.831506 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.831519 kubelet[3072]: W0913 01:21:51.831518 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.831613 kubelet[3072]: E0913 01:21:51.831529 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.831748 kubelet[3072]: E0913 01:21:51.831736 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.831748 kubelet[3072]: W0913 01:21:51.831746 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.831855 kubelet[3072]: E0913 01:21:51.831757 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.831931 kubelet[3072]: E0913 01:21:51.831919 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.831931 kubelet[3072]: W0913 01:21:51.831929 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.832021 kubelet[3072]: E0913 01:21:51.831939 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.832138 kubelet[3072]: E0913 01:21:51.832128 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.832138 kubelet[3072]: W0913 01:21:51.832138 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.832234 kubelet[3072]: E0913 01:21:51.832148 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.832328 kubelet[3072]: E0913 01:21:51.832316 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.832328 kubelet[3072]: W0913 01:21:51.832327 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.832420 kubelet[3072]: E0913 01:21:51.832337 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.832498 kubelet[3072]: E0913 01:21:51.832488 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.832549 kubelet[3072]: W0913 01:21:51.832498 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.832549 kubelet[3072]: E0913 01:21:51.832508 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.832682 kubelet[3072]: E0913 01:21:51.832672 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.832682 kubelet[3072]: W0913 01:21:51.832682 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.832764 kubelet[3072]: E0913 01:21:51.832691 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.832878 kubelet[3072]: E0913 01:21:51.832867 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.832926 kubelet[3072]: W0913 01:21:51.832877 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.832926 kubelet[3072]: E0913 01:21:51.832887 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.833064 kubelet[3072]: E0913 01:21:51.833053 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.833110 kubelet[3072]: W0913 01:21:51.833063 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.833110 kubelet[3072]: E0913 01:21:51.833074 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.852603 kubelet[3072]: E0913 01:21:51.852552 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.852603 kubelet[3072]: W0913 01:21:51.852571 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.852603 kubelet[3072]: E0913 01:21:51.852589 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.852905 kubelet[3072]: E0913 01:21:51.852887 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.852905 kubelet[3072]: W0913 01:21:51.852905 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.852988 kubelet[3072]: E0913 01:21:51.852922 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.853182 kubelet[3072]: E0913 01:21:51.853166 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.853182 kubelet[3072]: W0913 01:21:51.853179 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.853346 kubelet[3072]: E0913 01:21:51.853195 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.853507 kubelet[3072]: E0913 01:21:51.853491 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.853562 kubelet[3072]: W0913 01:21:51.853507 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.853562 kubelet[3072]: E0913 01:21:51.853524 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.853770 kubelet[3072]: E0913 01:21:51.853755 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.853845 kubelet[3072]: W0913 01:21:51.853769 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.853845 kubelet[3072]: E0913 01:21:51.853787 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.854020 kubelet[3072]: E0913 01:21:51.853999 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.854079 kubelet[3072]: W0913 01:21:51.854020 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.854079 kubelet[3072]: E0913 01:21:51.854056 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.854280 kubelet[3072]: E0913 01:21:51.854264 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.854280 kubelet[3072]: W0913 01:21:51.854278 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.854378 kubelet[3072]: E0913 01:21:51.854310 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.854473 kubelet[3072]: E0913 01:21:51.854462 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.854517 kubelet[3072]: W0913 01:21:51.854473 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.854517 kubelet[3072]: E0913 01:21:51.854503 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.854740 kubelet[3072]: E0913 01:21:51.854728 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.854793 kubelet[3072]: W0913 01:21:51.854740 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.854793 kubelet[3072]: E0913 01:21:51.854756 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.855044 kubelet[3072]: E0913 01:21:51.855029 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.855044 kubelet[3072]: W0913 01:21:51.855043 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.855125 kubelet[3072]: E0913 01:21:51.855060 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.855250 kubelet[3072]: E0913 01:21:51.855233 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.855306 kubelet[3072]: W0913 01:21:51.855251 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.855306 kubelet[3072]: E0913 01:21:51.855270 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.855507 kubelet[3072]: E0913 01:21:51.855495 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.855507 kubelet[3072]: W0913 01:21:51.855506 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.855585 kubelet[3072]: E0913 01:21:51.855522 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.855843 kubelet[3072]: E0913 01:21:51.855827 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.855843 kubelet[3072]: W0913 01:21:51.855839 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.855995 kubelet[3072]: E0913 01:21:51.855855 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.856067 kubelet[3072]: E0913 01:21:51.856047 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.856067 kubelet[3072]: W0913 01:21:51.856058 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.856206 kubelet[3072]: E0913 01:21:51.856072 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.856283 kubelet[3072]: E0913 01:21:51.856270 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.856363 kubelet[3072]: W0913 01:21:51.856298 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.856435 kubelet[3072]: E0913 01:21:51.856351 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.856513 kubelet[3072]: E0913 01:21:51.856465 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.856513 kubelet[3072]: W0913 01:21:51.856475 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.856641 kubelet[3072]: E0913 01:21:51.856503 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.856708 kubelet[3072]: E0913 01:21:51.856675 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.856708 kubelet[3072]: W0913 01:21:51.856686 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.856708 kubelet[3072]: E0913 01:21:51.856700 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:51.856979 kubelet[3072]: E0913 01:21:51.856964 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:51.857038 kubelet[3072]: W0913 01:21:51.856980 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:51.857038 kubelet[3072]: E0913 01:21:51.856998 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.769050 kubelet[3072]: I0913 01:21:52.768989 3072 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 01:21:52.840223 kubelet[3072]: E0913 01:21:52.840137 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.840223 kubelet[3072]: W0913 01:21:52.840178 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.840223 kubelet[3072]: E0913 01:21:52.840217 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.840869 kubelet[3072]: E0913 01:21:52.840824 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.840869 kubelet[3072]: W0913 01:21:52.840860 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.841108 kubelet[3072]: E0913 01:21:52.840892 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.841471 kubelet[3072]: E0913 01:21:52.841422 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.841471 kubelet[3072]: W0913 01:21:52.841453 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.841719 kubelet[3072]: E0913 01:21:52.841497 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.842086 kubelet[3072]: E0913 01:21:52.842028 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.842086 kubelet[3072]: W0913 01:21:52.842070 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.842430 kubelet[3072]: E0913 01:21:52.842109 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.842703 kubelet[3072]: E0913 01:21:52.842654 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.842703 kubelet[3072]: W0913 01:21:52.842683 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.842890 kubelet[3072]: E0913 01:21:52.842724 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.843361 kubelet[3072]: E0913 01:21:52.843296 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.843361 kubelet[3072]: W0913 01:21:52.843334 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.843589 kubelet[3072]: E0913 01:21:52.843375 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.843916 kubelet[3072]: E0913 01:21:52.843886 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.844007 kubelet[3072]: W0913 01:21:52.843922 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.844007 kubelet[3072]: E0913 01:21:52.843964 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.844594 kubelet[3072]: E0913 01:21:52.844539 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.844594 kubelet[3072]: W0913 01:21:52.844578 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.844934 kubelet[3072]: E0913 01:21:52.844621 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.845161 kubelet[3072]: E0913 01:21:52.845126 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.845328 kubelet[3072]: W0913 01:21:52.845168 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.845328 kubelet[3072]: E0913 01:21:52.845210 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.845789 kubelet[3072]: E0913 01:21:52.845715 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.845789 kubelet[3072]: W0913 01:21:52.845741 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.845789 kubelet[3072]: E0913 01:21:52.845768 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.846212 kubelet[3072]: E0913 01:21:52.846184 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.846212 kubelet[3072]: W0913 01:21:52.846210 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.846423 kubelet[3072]: E0913 01:21:52.846236 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.846850 kubelet[3072]: E0913 01:21:52.846782 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.846850 kubelet[3072]: W0913 01:21:52.846809 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.846850 kubelet[3072]: E0913 01:21:52.846834 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.847382 kubelet[3072]: E0913 01:21:52.847306 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.847382 kubelet[3072]: W0913 01:21:52.847329 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.847382 kubelet[3072]: E0913 01:21:52.847351 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.847851 kubelet[3072]: E0913 01:21:52.847800 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.847851 kubelet[3072]: W0913 01:21:52.847825 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.847851 kubelet[3072]: E0913 01:21:52.847848 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.848404 kubelet[3072]: E0913 01:21:52.848369 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.848404 kubelet[3072]: W0913 01:21:52.848395 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.848607 kubelet[3072]: E0913 01:21:52.848419 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.860923 kubelet[3072]: E0913 01:21:52.860848 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.860923 kubelet[3072]: W0913 01:21:52.860882 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.860923 kubelet[3072]: E0913 01:21:52.860914 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.861595 kubelet[3072]: E0913 01:21:52.861508 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.861595 kubelet[3072]: W0913 01:21:52.861536 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.861595 kubelet[3072]: E0913 01:21:52.861570 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.862185 kubelet[3072]: E0913 01:21:52.862097 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.862185 kubelet[3072]: W0913 01:21:52.862124 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.862185 kubelet[3072]: E0913 01:21:52.862156 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.862833 kubelet[3072]: E0913 01:21:52.862757 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.862833 kubelet[3072]: W0913 01:21:52.862791 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.862833 kubelet[3072]: E0913 01:21:52.862828 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.863439 kubelet[3072]: E0913 01:21:52.863355 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.863439 kubelet[3072]: W0913 01:21:52.863383 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.863679 kubelet[3072]: E0913 01:21:52.863487 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.863947 kubelet[3072]: E0913 01:21:52.863863 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.863947 kubelet[3072]: W0913 01:21:52.863888 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.864188 kubelet[3072]: E0913 01:21:52.863954 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.864423 kubelet[3072]: E0913 01:21:52.864342 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.864423 kubelet[3072]: W0913 01:21:52.864368 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.864666 kubelet[3072]: E0913 01:21:52.864426 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.864898 kubelet[3072]: E0913 01:21:52.864844 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.864898 kubelet[3072]: W0913 01:21:52.864870 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.865091 kubelet[3072]: E0913 01:21:52.864902 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.865502 kubelet[3072]: E0913 01:21:52.865453 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.865502 kubelet[3072]: W0913 01:21:52.865478 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.865697 kubelet[3072]: E0913 01:21:52.865510 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.866094 kubelet[3072]: E0913 01:21:52.866045 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.866094 kubelet[3072]: W0913 01:21:52.866076 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.866315 kubelet[3072]: E0913 01:21:52.866113 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.866669 kubelet[3072]: E0913 01:21:52.866620 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.866669 kubelet[3072]: W0913 01:21:52.866648 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.866874 kubelet[3072]: E0913 01:21:52.866745 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.867198 kubelet[3072]: E0913 01:21:52.867163 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.867347 kubelet[3072]: W0913 01:21:52.867202 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.867347 kubelet[3072]: E0913 01:21:52.867294 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.867769 kubelet[3072]: E0913 01:21:52.867714 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.867769 kubelet[3072]: W0913 01:21:52.867742 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.867968 kubelet[3072]: E0913 01:21:52.867821 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.868323 kubelet[3072]: E0913 01:21:52.868289 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.868323 kubelet[3072]: W0913 01:21:52.868321 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.868513 kubelet[3072]: E0913 01:21:52.868353 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.869011 kubelet[3072]: E0913 01:21:52.868956 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.869011 kubelet[3072]: W0913 01:21:52.868990 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.869196 kubelet[3072]: E0913 01:21:52.869029 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.869584 kubelet[3072]: E0913 01:21:52.869554 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.869584 kubelet[3072]: W0913 01:21:52.869583 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.869777 kubelet[3072]: E0913 01:21:52.869617 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.870369 kubelet[3072]: E0913 01:21:52.870279 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.870369 kubelet[3072]: W0913 01:21:52.870324 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.870369 kubelet[3072]: E0913 01:21:52.870365 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:52.871082 kubelet[3072]: E0913 01:21:52.871015 3072 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 13 01:21:52.871082 kubelet[3072]: W0913 01:21:52.871042 3072 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 13 01:21:52.871082 kubelet[3072]: E0913 01:21:52.871070 3072 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 13 01:21:53.281116 containerd[1819]: time="2025-09-13T01:21:53.281066041Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:53.281364 containerd[1819]: time="2025-09-13T01:21:53.281235318Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4446660" Sep 13 01:21:53.281693 containerd[1819]: time="2025-09-13T01:21:53.281651556Z" level=info msg="ImageCreate event name:\"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:53.282658 containerd[1819]: time="2025-09-13T01:21:53.282618425Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:53.283053 containerd[1819]: time="2025-09-13T01:21:53.283009654Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5939323\" in 1.642263428s" Sep 13 01:21:53.283053 containerd[1819]: time="2025-09-13T01:21:53.283027320Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\"" Sep 13 01:21:53.284371 containerd[1819]: time="2025-09-13T01:21:53.284355122Z" level=info msg="CreateContainer within sandbox \"36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 13 01:21:53.289221 containerd[1819]: time="2025-09-13T01:21:53.289175735Z" level=info msg="CreateContainer within sandbox \"36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640\"" Sep 13 01:21:53.289489 containerd[1819]: time="2025-09-13T01:21:53.289439108Z" level=info msg="StartContainer for \"3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640\"" Sep 13 01:21:53.314414 systemd[1]: Started cri-containerd-3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640.scope - libcontainer container 3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640. Sep 13 01:21:53.329266 containerd[1819]: time="2025-09-13T01:21:53.329150587Z" level=info msg="StartContainer for \"3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640\" returns successfully" Sep 13 01:21:53.335042 systemd[1]: cri-containerd-3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640.scope: Deactivated successfully. Sep 13 01:21:53.648373 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640-rootfs.mount: Deactivated successfully. Sep 13 01:21:53.721393 kubelet[3072]: E0913 01:21:53.721280 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-jqdjm" podUID="f4d72f52-2d93-4ed1-9fa8-3baa96684daf" Sep 13 01:21:53.783065 containerd[1819]: time="2025-09-13T01:21:53.783033666Z" level=info msg="shim disconnected" id=3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640 namespace=k8s.io Sep 13 01:21:53.783065 containerd[1819]: time="2025-09-13T01:21:53.783062883Z" level=warning msg="cleaning up after shim disconnected" id=3bcdd14d8d04fa89f41ffd1f86f7e6572c6d610a168983337e4a58b0bb8cc640 namespace=k8s.io Sep 13 01:21:53.783065 containerd[1819]: time="2025-09-13T01:21:53.783068850Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:21:54.782803 containerd[1819]: time="2025-09-13T01:21:54.782735235Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 13 01:21:55.721235 kubelet[3072]: E0913 01:21:55.721139 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-jqdjm" podUID="f4d72f52-2d93-4ed1-9fa8-3baa96684daf" Sep 13 01:21:57.720570 kubelet[3072]: E0913 01:21:57.720546 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-jqdjm" podUID="f4d72f52-2d93-4ed1-9fa8-3baa96684daf" Sep 13 01:21:57.772902 containerd[1819]: time="2025-09-13T01:21:57.772877093Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:57.773178 containerd[1819]: time="2025-09-13T01:21:57.773074589Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=70440613" Sep 13 01:21:57.773450 containerd[1819]: time="2025-09-13T01:21:57.773437750Z" level=info msg="ImageCreate event name:\"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:57.774536 containerd[1819]: time="2025-09-13T01:21:57.774523827Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:21:57.774966 containerd[1819]: time="2025-09-13T01:21:57.774949459Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"71933316\" in 2.99214695s" Sep 13 01:21:57.774999 containerd[1819]: time="2025-09-13T01:21:57.774968933Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\"" Sep 13 01:21:57.775919 containerd[1819]: time="2025-09-13T01:21:57.775907400Z" level=info msg="CreateContainer within sandbox \"36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 13 01:21:57.781267 containerd[1819]: time="2025-09-13T01:21:57.781205959Z" level=info msg="CreateContainer within sandbox \"36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d\"" Sep 13 01:21:57.781496 containerd[1819]: time="2025-09-13T01:21:57.781452574Z" level=info msg="StartContainer for \"e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d\"" Sep 13 01:21:57.814704 systemd[1]: Started cri-containerd-e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d.scope - libcontainer container e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d. Sep 13 01:21:57.869779 containerd[1819]: time="2025-09-13T01:21:57.869715961Z" level=info msg="StartContainer for \"e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d\" returns successfully" Sep 13 01:21:58.484851 systemd[1]: cri-containerd-e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d.scope: Deactivated successfully. Sep 13 01:21:58.556618 kubelet[3072]: I0913 01:21:58.556522 3072 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 13 01:21:58.615155 systemd[1]: Created slice kubepods-besteffort-pod198f178f_c689_48cb_b5e4_012f47ac61f3.slice - libcontainer container kubepods-besteffort-pod198f178f_c689_48cb_b5e4_012f47ac61f3.slice. Sep 13 01:21:58.617135 kubelet[3072]: I0913 01:21:58.616560 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzcn\" (UniqueName: \"kubernetes.io/projected/198f178f-c689-48cb-b5e4-012f47ac61f3-kube-api-access-zqzcn\") pod \"calico-kube-controllers-689db9d4f-r55km\" (UID: \"198f178f-c689-48cb-b5e4-012f47ac61f3\") " pod="calico-system/calico-kube-controllers-689db9d4f-r55km" Sep 13 01:21:58.617135 kubelet[3072]: I0913 01:21:58.616683 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/198f178f-c689-48cb-b5e4-012f47ac61f3-tigera-ca-bundle\") pod \"calico-kube-controllers-689db9d4f-r55km\" (UID: \"198f178f-c689-48cb-b5e4-012f47ac61f3\") " pod="calico-system/calico-kube-controllers-689db9d4f-r55km" Sep 13 01:21:58.629409 systemd[1]: Created slice kubepods-burstable-pod465526ae_5203_429b_8157_1675929b92b0.slice - libcontainer container kubepods-burstable-pod465526ae_5203_429b_8157_1675929b92b0.slice. Sep 13 01:21:58.633463 systemd[1]: Created slice kubepods-burstable-pod0db74307_721a_4439_bd8d_982b96aa7db8.slice - libcontainer container kubepods-burstable-pod0db74307_721a_4439_bd8d_982b96aa7db8.slice. Sep 13 01:21:58.636936 systemd[1]: Created slice kubepods-besteffort-pod1e417130_1b3b_4716_84aa_d920f0d71e2e.slice - libcontainer container kubepods-besteffort-pod1e417130_1b3b_4716_84aa_d920f0d71e2e.slice. Sep 13 01:21:58.640593 systemd[1]: Created slice kubepods-besteffort-pod3999b298_1754_45b1_95dc_2f560b55d1e4.slice - libcontainer container kubepods-besteffort-pod3999b298_1754_45b1_95dc_2f560b55d1e4.slice. Sep 13 01:21:58.643745 systemd[1]: Created slice kubepods-besteffort-pod971bad55_1380_4bfd_8af5_2cbef17292ea.slice - libcontainer container kubepods-besteffort-pod971bad55_1380_4bfd_8af5_2cbef17292ea.slice. Sep 13 01:21:58.646457 systemd[1]: Created slice kubepods-besteffort-pod878b3bf4_9638_46e7_9a71_02b8d2591f7f.slice - libcontainer container kubepods-besteffort-pod878b3bf4_9638_46e7_9a71_02b8d2591f7f.slice. Sep 13 01:21:58.717467 kubelet[3072]: I0913 01:21:58.717401 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctrsl\" (UniqueName: \"kubernetes.io/projected/971bad55-1380-4bfd-8af5-2cbef17292ea-kube-api-access-ctrsl\") pod \"whisker-85749946d-xvsn5\" (UID: \"971bad55-1380-4bfd-8af5-2cbef17292ea\") " pod="calico-system/whisker-85749946d-xvsn5" Sep 13 01:21:58.717467 kubelet[3072]: I0913 01:21:58.717447 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3999b298-1754-45b1-95dc-2f560b55d1e4-config\") pod \"goldmane-54d579b49d-sfpqb\" (UID: \"3999b298-1754-45b1-95dc-2f560b55d1e4\") " pod="calico-system/goldmane-54d579b49d-sfpqb" Sep 13 01:21:58.717467 kubelet[3072]: I0913 01:21:58.717472 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3999b298-1754-45b1-95dc-2f560b55d1e4-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-sfpqb\" (UID: \"3999b298-1754-45b1-95dc-2f560b55d1e4\") " pod="calico-system/goldmane-54d579b49d-sfpqb" Sep 13 01:21:58.727597 kubelet[3072]: I0913 01:21:58.717494 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/1e417130-1b3b-4716-84aa-d920f0d71e2e-calico-apiserver-certs\") pod \"calico-apiserver-db75486c4-rqw44\" (UID: \"1e417130-1b3b-4716-84aa-d920f0d71e2e\") " pod="calico-apiserver/calico-apiserver-db75486c4-rqw44" Sep 13 01:21:58.727597 kubelet[3072]: I0913 01:21:58.717516 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0db74307-721a-4439-bd8d-982b96aa7db8-config-volume\") pod \"coredns-668d6bf9bc-zkzzg\" (UID: \"0db74307-721a-4439-bd8d-982b96aa7db8\") " pod="kube-system/coredns-668d6bf9bc-zkzzg" Sep 13 01:21:58.727597 kubelet[3072]: I0913 01:21:58.717572 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-backend-key-pair\") pod \"whisker-85749946d-xvsn5\" (UID: \"971bad55-1380-4bfd-8af5-2cbef17292ea\") " pod="calico-system/whisker-85749946d-xvsn5" Sep 13 01:21:58.727597 kubelet[3072]: I0913 01:21:58.717648 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk24w\" (UniqueName: \"kubernetes.io/projected/878b3bf4-9638-46e7-9a71-02b8d2591f7f-kube-api-access-xk24w\") pod \"calico-apiserver-db75486c4-sm7wq\" (UID: \"878b3bf4-9638-46e7-9a71-02b8d2591f7f\") " pod="calico-apiserver/calico-apiserver-db75486c4-sm7wq" Sep 13 01:21:58.727597 kubelet[3072]: I0913 01:21:58.717697 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/3999b298-1754-45b1-95dc-2f560b55d1e4-goldmane-key-pair\") pod \"goldmane-54d579b49d-sfpqb\" (UID: \"3999b298-1754-45b1-95dc-2f560b55d1e4\") " pod="calico-system/goldmane-54d579b49d-sfpqb" Sep 13 01:21:58.727881 kubelet[3072]: I0913 01:21:58.717719 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hjl6\" (UniqueName: \"kubernetes.io/projected/3999b298-1754-45b1-95dc-2f560b55d1e4-kube-api-access-6hjl6\") pod \"goldmane-54d579b49d-sfpqb\" (UID: \"3999b298-1754-45b1-95dc-2f560b55d1e4\") " pod="calico-system/goldmane-54d579b49d-sfpqb" Sep 13 01:21:58.727881 kubelet[3072]: I0913 01:21:58.717755 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk6q8\" (UniqueName: \"kubernetes.io/projected/1e417130-1b3b-4716-84aa-d920f0d71e2e-kube-api-access-bk6q8\") pod \"calico-apiserver-db75486c4-rqw44\" (UID: \"1e417130-1b3b-4716-84aa-d920f0d71e2e\") " pod="calico-apiserver/calico-apiserver-db75486c4-rqw44" Sep 13 01:21:58.727881 kubelet[3072]: I0913 01:21:58.717776 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvx46\" (UniqueName: \"kubernetes.io/projected/0db74307-721a-4439-bd8d-982b96aa7db8-kube-api-access-lvx46\") pod \"coredns-668d6bf9bc-zkzzg\" (UID: \"0db74307-721a-4439-bd8d-982b96aa7db8\") " pod="kube-system/coredns-668d6bf9bc-zkzzg" Sep 13 01:21:58.727881 kubelet[3072]: I0913 01:21:58.717807 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/878b3bf4-9638-46e7-9a71-02b8d2591f7f-calico-apiserver-certs\") pod \"calico-apiserver-db75486c4-sm7wq\" (UID: \"878b3bf4-9638-46e7-9a71-02b8d2591f7f\") " pod="calico-apiserver/calico-apiserver-db75486c4-sm7wq" Sep 13 01:21:58.727881 kubelet[3072]: I0913 01:21:58.717829 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/465526ae-5203-429b-8157-1675929b92b0-config-volume\") pod \"coredns-668d6bf9bc-5hgc5\" (UID: \"465526ae-5203-429b-8157-1675929b92b0\") " pod="kube-system/coredns-668d6bf9bc-5hgc5" Sep 13 01:21:58.728163 kubelet[3072]: I0913 01:21:58.717850 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-ca-bundle\") pod \"whisker-85749946d-xvsn5\" (UID: \"971bad55-1380-4bfd-8af5-2cbef17292ea\") " pod="calico-system/whisker-85749946d-xvsn5" Sep 13 01:21:58.728163 kubelet[3072]: I0913 01:21:58.717899 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl45b\" (UniqueName: \"kubernetes.io/projected/465526ae-5203-429b-8157-1675929b92b0-kube-api-access-gl45b\") pod \"coredns-668d6bf9bc-5hgc5\" (UID: \"465526ae-5203-429b-8157-1675929b92b0\") " pod="kube-system/coredns-668d6bf9bc-5hgc5" Sep 13 01:21:58.788346 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d-rootfs.mount: Deactivated successfully. Sep 13 01:21:58.902720 containerd[1819]: time="2025-09-13T01:21:58.902661834Z" level=info msg="shim disconnected" id=e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d namespace=k8s.io Sep 13 01:21:58.902720 containerd[1819]: time="2025-09-13T01:21:58.902697346Z" level=warning msg="cleaning up after shim disconnected" id=e66c6e33bc31b835dfd24b9873d73688be06560fd9f7eea95700d34fbfb2d47d namespace=k8s.io Sep 13 01:21:58.902720 containerd[1819]: time="2025-09-13T01:21:58.902703993Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:21:58.929691 containerd[1819]: time="2025-09-13T01:21:58.929612664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-689db9d4f-r55km,Uid:198f178f-c689-48cb-b5e4-012f47ac61f3,Namespace:calico-system,Attempt:0,}" Sep 13 01:21:58.932036 containerd[1819]: time="2025-09-13T01:21:58.931998515Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5hgc5,Uid:465526ae-5203-429b-8157-1675929b92b0,Namespace:kube-system,Attempt:0,}" Sep 13 01:21:58.935355 containerd[1819]: time="2025-09-13T01:21:58.935320936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-zkzzg,Uid:0db74307-721a-4439-bd8d-982b96aa7db8,Namespace:kube-system,Attempt:0,}" Sep 13 01:21:58.938765 containerd[1819]: time="2025-09-13T01:21:58.938744551Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-rqw44,Uid:1e417130-1b3b-4716-84aa-d920f0d71e2e,Namespace:calico-apiserver,Attempt:0,}" Sep 13 01:21:58.943332 containerd[1819]: time="2025-09-13T01:21:58.943310204Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-sfpqb,Uid:3999b298-1754-45b1-95dc-2f560b55d1e4,Namespace:calico-system,Attempt:0,}" Sep 13 01:21:58.945777 containerd[1819]: time="2025-09-13T01:21:58.945754499Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-85749946d-xvsn5,Uid:971bad55-1380-4bfd-8af5-2cbef17292ea,Namespace:calico-system,Attempt:0,}" Sep 13 01:21:58.948197 containerd[1819]: time="2025-09-13T01:21:58.948172428Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-sm7wq,Uid:878b3bf4-9638-46e7-9a71-02b8d2591f7f,Namespace:calico-apiserver,Attempt:0,}" Sep 13 01:21:58.963462 containerd[1819]: time="2025-09-13T01:21:58.963408544Z" level=error msg="Failed to destroy network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.963727 containerd[1819]: time="2025-09-13T01:21:58.963707896Z" level=error msg="encountered an error cleaning up failed sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.963770 containerd[1819]: time="2025-09-13T01:21:58.963753724Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-689db9d4f-r55km,Uid:198f178f-c689-48cb-b5e4-012f47ac61f3,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.963947 kubelet[3072]: E0913 01:21:58.963920 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.964021 kubelet[3072]: E0913 01:21:58.963983 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-689db9d4f-r55km" Sep 13 01:21:58.964065 kubelet[3072]: E0913 01:21:58.964018 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-689db9d4f-r55km" Sep 13 01:21:58.964096 kubelet[3072]: E0913 01:21:58.964059 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-689db9d4f-r55km_calico-system(198f178f-c689-48cb-b5e4-012f47ac61f3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-689db9d4f-r55km_calico-system(198f178f-c689-48cb-b5e4-012f47ac61f3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-689db9d4f-r55km" podUID="198f178f-c689-48cb-b5e4-012f47ac61f3" Sep 13 01:21:58.964745 containerd[1819]: time="2025-09-13T01:21:58.964723123Z" level=error msg="Failed to destroy network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.964915 containerd[1819]: time="2025-09-13T01:21:58.964899180Z" level=error msg="encountered an error cleaning up failed sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.964948 containerd[1819]: time="2025-09-13T01:21:58.964938341Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5hgc5,Uid:465526ae-5203-429b-8157-1675929b92b0,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.965073 kubelet[3072]: E0913 01:21:58.965047 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.965108 kubelet[3072]: E0913 01:21:58.965095 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-5hgc5" Sep 13 01:21:58.965130 kubelet[3072]: E0913 01:21:58.965117 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-5hgc5" Sep 13 01:21:58.965171 kubelet[3072]: E0913 01:21:58.965154 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-5hgc5_kube-system(465526ae-5203-429b-8157-1675929b92b0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-5hgc5_kube-system(465526ae-5203-429b-8157-1675929b92b0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-5hgc5" podUID="465526ae-5203-429b-8157-1675929b92b0" Sep 13 01:21:58.966131 containerd[1819]: time="2025-09-13T01:21:58.966112013Z" level=error msg="Failed to destroy network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.966305 containerd[1819]: time="2025-09-13T01:21:58.966292617Z" level=error msg="encountered an error cleaning up failed sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.966333 containerd[1819]: time="2025-09-13T01:21:58.966317972Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-zkzzg,Uid:0db74307-721a-4439-bd8d-982b96aa7db8,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.966470 kubelet[3072]: E0913 01:21:58.966452 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.966505 kubelet[3072]: E0913 01:21:58.966485 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-zkzzg" Sep 13 01:21:58.966505 kubelet[3072]: E0913 01:21:58.966499 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-zkzzg" Sep 13 01:21:58.966542 kubelet[3072]: E0913 01:21:58.966521 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-zkzzg_kube-system(0db74307-721a-4439-bd8d-982b96aa7db8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-zkzzg_kube-system(0db74307-721a-4439-bd8d-982b96aa7db8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-zkzzg" podUID="0db74307-721a-4439-bd8d-982b96aa7db8" Sep 13 01:21:58.968787 containerd[1819]: time="2025-09-13T01:21:58.968758998Z" level=error msg="Failed to destroy network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.968951 containerd[1819]: time="2025-09-13T01:21:58.968938304Z" level=error msg="encountered an error cleaning up failed sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.968992 containerd[1819]: time="2025-09-13T01:21:58.968966356Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-rqw44,Uid:1e417130-1b3b-4716-84aa-d920f0d71e2e,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.969081 kubelet[3072]: E0913 01:21:58.969065 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.969127 kubelet[3072]: E0913 01:21:58.969095 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-db75486c4-rqw44" Sep 13 01:21:58.969127 kubelet[3072]: E0913 01:21:58.969109 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-db75486c4-rqw44" Sep 13 01:21:58.969196 kubelet[3072]: E0913 01:21:58.969153 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-db75486c4-rqw44_calico-apiserver(1e417130-1b3b-4716-84aa-d920f0d71e2e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-db75486c4-rqw44_calico-apiserver(1e417130-1b3b-4716-84aa-d920f0d71e2e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-db75486c4-rqw44" podUID="1e417130-1b3b-4716-84aa-d920f0d71e2e" Sep 13 01:21:58.974909 containerd[1819]: time="2025-09-13T01:21:58.974879781Z" level=error msg="Failed to destroy network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.975067 containerd[1819]: time="2025-09-13T01:21:58.975053347Z" level=error msg="encountered an error cleaning up failed sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.975096 containerd[1819]: time="2025-09-13T01:21:58.975085364Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-sfpqb,Uid:3999b298-1754-45b1-95dc-2f560b55d1e4,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.975225 kubelet[3072]: E0913 01:21:58.975203 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.975302 kubelet[3072]: E0913 01:21:58.975247 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-sfpqb" Sep 13 01:21:58.975302 kubelet[3072]: E0913 01:21:58.975269 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-sfpqb" Sep 13 01:21:58.975364 kubelet[3072]: E0913 01:21:58.975307 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-sfpqb_calico-system(3999b298-1754-45b1-95dc-2f560b55d1e4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-sfpqb_calico-system(3999b298-1754-45b1-95dc-2f560b55d1e4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-sfpqb" podUID="3999b298-1754-45b1-95dc-2f560b55d1e4" Sep 13 01:21:58.982897 containerd[1819]: time="2025-09-13T01:21:58.982865737Z" level=error msg="Failed to destroy network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.983074 containerd[1819]: time="2025-09-13T01:21:58.983060246Z" level=error msg="encountered an error cleaning up failed sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.983109 containerd[1819]: time="2025-09-13T01:21:58.983096177Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-85749946d-xvsn5,Uid:971bad55-1380-4bfd-8af5-2cbef17292ea,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.983262 kubelet[3072]: E0913 01:21:58.983229 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.983323 kubelet[3072]: E0913 01:21:58.983284 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-85749946d-xvsn5" Sep 13 01:21:58.983323 kubelet[3072]: E0913 01:21:58.983303 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-85749946d-xvsn5" Sep 13 01:21:58.983390 kubelet[3072]: E0913 01:21:58.983340 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-85749946d-xvsn5_calico-system(971bad55-1380-4bfd-8af5-2cbef17292ea)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-85749946d-xvsn5_calico-system(971bad55-1380-4bfd-8af5-2cbef17292ea)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-85749946d-xvsn5" podUID="971bad55-1380-4bfd-8af5-2cbef17292ea" Sep 13 01:21:58.998083 containerd[1819]: time="2025-09-13T01:21:58.998054154Z" level=error msg="Failed to destroy network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.998329 containerd[1819]: time="2025-09-13T01:21:58.998240938Z" level=error msg="encountered an error cleaning up failed sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.998329 containerd[1819]: time="2025-09-13T01:21:58.998274372Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-sm7wq,Uid:878b3bf4-9638-46e7-9a71-02b8d2591f7f,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.998427 kubelet[3072]: E0913 01:21:58.998405 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:58.998474 kubelet[3072]: E0913 01:21:58.998444 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-db75486c4-sm7wq" Sep 13 01:21:58.998474 kubelet[3072]: E0913 01:21:58.998462 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-db75486c4-sm7wq" Sep 13 01:21:58.998534 kubelet[3072]: E0913 01:21:58.998494 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-db75486c4-sm7wq_calico-apiserver(878b3bf4-9638-46e7-9a71-02b8d2591f7f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-db75486c4-sm7wq_calico-apiserver(878b3bf4-9638-46e7-9a71-02b8d2591f7f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-db75486c4-sm7wq" podUID="878b3bf4-9638-46e7-9a71-02b8d2591f7f" Sep 13 01:21:59.736390 systemd[1]: Created slice kubepods-besteffort-podf4d72f52_2d93_4ed1_9fa8_3baa96684daf.slice - libcontainer container kubepods-besteffort-podf4d72f52_2d93_4ed1_9fa8_3baa96684daf.slice. Sep 13 01:21:59.741713 containerd[1819]: time="2025-09-13T01:21:59.741606307Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-jqdjm,Uid:f4d72f52-2d93-4ed1-9fa8-3baa96684daf,Namespace:calico-system,Attempt:0,}" Sep 13 01:21:59.769445 containerd[1819]: time="2025-09-13T01:21:59.769416515Z" level=error msg="Failed to destroy network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.769623 containerd[1819]: time="2025-09-13T01:21:59.769609293Z" level=error msg="encountered an error cleaning up failed sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.769652 containerd[1819]: time="2025-09-13T01:21:59.769642549Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-jqdjm,Uid:f4d72f52-2d93-4ed1-9fa8-3baa96684daf,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.769817 kubelet[3072]: E0913 01:21:59.769766 3072 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.769817 kubelet[3072]: E0913 01:21:59.769803 3072 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-jqdjm" Sep 13 01:21:59.770028 kubelet[3072]: E0913 01:21:59.769817 3072 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-jqdjm" Sep 13 01:21:59.770028 kubelet[3072]: E0913 01:21:59.769848 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-jqdjm_calico-system(f4d72f52-2d93-4ed1-9fa8-3baa96684daf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-jqdjm_calico-system(f4d72f52-2d93-4ed1-9fa8-3baa96684daf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-jqdjm" podUID="f4d72f52-2d93-4ed1-9fa8-3baa96684daf" Sep 13 01:21:59.782587 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272-shm.mount: Deactivated successfully. Sep 13 01:21:59.795932 kubelet[3072]: I0913 01:21:59.795917 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:21:59.796446 containerd[1819]: time="2025-09-13T01:21:59.796423175Z" level=info msg="StopPodSandbox for \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\"" Sep 13 01:21:59.796501 kubelet[3072]: I0913 01:21:59.796455 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:21:59.796669 containerd[1819]: time="2025-09-13T01:21:59.796649288Z" level=info msg="Ensure that sandbox 0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6 in task-service has been cleanup successfully" Sep 13 01:21:59.796818 containerd[1819]: time="2025-09-13T01:21:59.796802082Z" level=info msg="StopPodSandbox for \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\"" Sep 13 01:21:59.796981 containerd[1819]: time="2025-09-13T01:21:59.796964678Z" level=info msg="Ensure that sandbox 6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba in task-service has been cleanup successfully" Sep 13 01:21:59.797167 kubelet[3072]: I0913 01:21:59.797153 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:21:59.797536 containerd[1819]: time="2025-09-13T01:21:59.797516738Z" level=info msg="StopPodSandbox for \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\"" Sep 13 01:21:59.797749 containerd[1819]: time="2025-09-13T01:21:59.797730066Z" level=info msg="Ensure that sandbox 9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861 in task-service has been cleanup successfully" Sep 13 01:21:59.797847 kubelet[3072]: I0913 01:21:59.797834 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:21:59.798271 containerd[1819]: time="2025-09-13T01:21:59.798242915Z" level=info msg="StopPodSandbox for \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\"" Sep 13 01:21:59.798823 containerd[1819]: time="2025-09-13T01:21:59.798785998Z" level=info msg="Ensure that sandbox 5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911 in task-service has been cleanup successfully" Sep 13 01:21:59.800009 kubelet[3072]: I0913 01:21:59.799982 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:21:59.800569 containerd[1819]: time="2025-09-13T01:21:59.800534685Z" level=info msg="StopPodSandbox for \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\"" Sep 13 01:21:59.800757 containerd[1819]: time="2025-09-13T01:21:59.800734744Z" level=info msg="Ensure that sandbox c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07 in task-service has been cleanup successfully" Sep 13 01:21:59.801428 kubelet[3072]: I0913 01:21:59.801405 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:21:59.801934 containerd[1819]: time="2025-09-13T01:21:59.801901810Z" level=info msg="StopPodSandbox for \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\"" Sep 13 01:21:59.802133 containerd[1819]: time="2025-09-13T01:21:59.802106677Z" level=info msg="Ensure that sandbox 08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b in task-service has been cleanup successfully" Sep 13 01:21:59.802357 kubelet[3072]: I0913 01:21:59.802338 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:21:59.802809 containerd[1819]: time="2025-09-13T01:21:59.802786101Z" level=info msg="StopPodSandbox for \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\"" Sep 13 01:21:59.802964 containerd[1819]: time="2025-09-13T01:21:59.802946058Z" level=info msg="Ensure that sandbox de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7 in task-service has been cleanup successfully" Sep 13 01:21:59.803085 kubelet[3072]: I0913 01:21:59.803067 3072 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:21:59.803520 containerd[1819]: time="2025-09-13T01:21:59.803492362Z" level=info msg="StopPodSandbox for \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\"" Sep 13 01:21:59.803708 containerd[1819]: time="2025-09-13T01:21:59.803688672Z" level=info msg="Ensure that sandbox 5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272 in task-service has been cleanup successfully" Sep 13 01:21:59.805389 containerd[1819]: time="2025-09-13T01:21:59.805362875Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 13 01:21:59.814929 containerd[1819]: time="2025-09-13T01:21:59.814885782Z" level=error msg="StopPodSandbox for \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\" failed" error="failed to destroy network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.815062 kubelet[3072]: E0913 01:21:59.815037 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:21:59.815135 kubelet[3072]: E0913 01:21:59.815086 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911"} Sep 13 01:21:59.815173 kubelet[3072]: E0913 01:21:59.815153 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1e417130-1b3b-4716-84aa-d920f0d71e2e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.815231 kubelet[3072]: E0913 01:21:59.815175 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1e417130-1b3b-4716-84aa-d920f0d71e2e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-db75486c4-rqw44" podUID="1e417130-1b3b-4716-84aa-d920f0d71e2e" Sep 13 01:21:59.816154 containerd[1819]: time="2025-09-13T01:21:59.816120909Z" level=error msg="StopPodSandbox for \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\" failed" error="failed to destroy network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.816238 containerd[1819]: time="2025-09-13T01:21:59.816121487Z" level=error msg="StopPodSandbox for \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\" failed" error="failed to destroy network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.816374 kubelet[3072]: E0913 01:21:59.816350 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:21:59.816439 kubelet[3072]: E0913 01:21:59.816387 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6"} Sep 13 01:21:59.816439 kubelet[3072]: E0913 01:21:59.816417 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f4d72f52-2d93-4ed1-9fa8-3baa96684daf\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.816439 kubelet[3072]: E0913 01:21:59.816350 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:21:59.816560 kubelet[3072]: E0913 01:21:59.816437 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f4d72f52-2d93-4ed1-9fa8-3baa96684daf\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-jqdjm" podUID="f4d72f52-2d93-4ed1-9fa8-3baa96684daf" Sep 13 01:21:59.816560 kubelet[3072]: E0913 01:21:59.816444 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba"} Sep 13 01:21:59.816560 kubelet[3072]: E0913 01:21:59.816473 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"465526ae-5203-429b-8157-1675929b92b0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.816560 kubelet[3072]: E0913 01:21:59.816489 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"465526ae-5203-429b-8157-1675929b92b0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-5hgc5" podUID="465526ae-5203-429b-8157-1675929b92b0" Sep 13 01:21:59.818743 containerd[1819]: time="2025-09-13T01:21:59.818714421Z" level=error msg="StopPodSandbox for \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\" failed" error="failed to destroy network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.818855 kubelet[3072]: E0913 01:21:59.818834 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:21:59.818895 kubelet[3072]: E0913 01:21:59.818864 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861"} Sep 13 01:21:59.818895 kubelet[3072]: E0913 01:21:59.818883 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"971bad55-1380-4bfd-8af5-2cbef17292ea\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.818960 kubelet[3072]: E0913 01:21:59.818896 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"971bad55-1380-4bfd-8af5-2cbef17292ea\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-85749946d-xvsn5" podUID="971bad55-1380-4bfd-8af5-2cbef17292ea" Sep 13 01:21:59.820133 containerd[1819]: time="2025-09-13T01:21:59.820110332Z" level=error msg="StopPodSandbox for \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\" failed" error="failed to destroy network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.820278 kubelet[3072]: E0913 01:21:59.820241 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:21:59.820317 kubelet[3072]: E0913 01:21:59.820289 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07"} Sep 13 01:21:59.820338 kubelet[3072]: E0913 01:21:59.820315 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"0db74307-721a-4439-bd8d-982b96aa7db8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.820372 kubelet[3072]: E0913 01:21:59.820337 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"0db74307-721a-4439-bd8d-982b96aa7db8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-zkzzg" podUID="0db74307-721a-4439-bd8d-982b96aa7db8" Sep 13 01:21:59.820558 containerd[1819]: time="2025-09-13T01:21:59.820538201Z" level=error msg="StopPodSandbox for \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\" failed" error="failed to destroy network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.820630 kubelet[3072]: E0913 01:21:59.820616 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:21:59.820655 kubelet[3072]: E0913 01:21:59.820635 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7"} Sep 13 01:21:59.820655 kubelet[3072]: E0913 01:21:59.820649 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"3999b298-1754-45b1-95dc-2f560b55d1e4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.820702 kubelet[3072]: E0913 01:21:59.820660 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"3999b298-1754-45b1-95dc-2f560b55d1e4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-sfpqb" podUID="3999b298-1754-45b1-95dc-2f560b55d1e4" Sep 13 01:21:59.821031 containerd[1819]: time="2025-09-13T01:21:59.821016078Z" level=error msg="StopPodSandbox for \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\" failed" error="failed to destroy network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.821085 kubelet[3072]: E0913 01:21:59.821073 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:21:59.821110 kubelet[3072]: E0913 01:21:59.821090 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b"} Sep 13 01:21:59.821110 kubelet[3072]: E0913 01:21:59.821105 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"878b3bf4-9638-46e7-9a71-02b8d2591f7f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.821155 kubelet[3072]: E0913 01:21:59.821115 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"878b3bf4-9638-46e7-9a71-02b8d2591f7f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-db75486c4-sm7wq" podUID="878b3bf4-9638-46e7-9a71-02b8d2591f7f" Sep 13 01:21:59.821833 containerd[1819]: time="2025-09-13T01:21:59.821818872Z" level=error msg="StopPodSandbox for \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\" failed" error="failed to destroy network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 13 01:21:59.821894 kubelet[3072]: E0913 01:21:59.821883 3072 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:21:59.821925 kubelet[3072]: E0913 01:21:59.821898 3072 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272"} Sep 13 01:21:59.821925 kubelet[3072]: E0913 01:21:59.821910 3072 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"198f178f-c689-48cb-b5e4-012f47ac61f3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 13 01:21:59.821972 kubelet[3072]: E0913 01:21:59.821923 3072 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"198f178f-c689-48cb-b5e4-012f47ac61f3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-689db9d4f-r55km" podUID="198f178f-c689-48cb-b5e4-012f47ac61f3" Sep 13 01:22:05.137033 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2076348271.mount: Deactivated successfully. Sep 13 01:22:05.162037 containerd[1819]: time="2025-09-13T01:22:05.161988000Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:05.162211 containerd[1819]: time="2025-09-13T01:22:05.162126975Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=157078339" Sep 13 01:22:05.162534 containerd[1819]: time="2025-09-13T01:22:05.162518270Z" level=info msg="ImageCreate event name:\"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:05.163403 containerd[1819]: time="2025-09-13T01:22:05.163389636Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:05.163777 containerd[1819]: time="2025-09-13T01:22:05.163764487Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"157078201\" in 5.3583751s" Sep 13 01:22:05.163804 containerd[1819]: time="2025-09-13T01:22:05.163780029Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\"" Sep 13 01:22:05.167062 containerd[1819]: time="2025-09-13T01:22:05.167039588Z" level=info msg="CreateContainer within sandbox \"36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 13 01:22:05.172770 containerd[1819]: time="2025-09-13T01:22:05.172750295Z" level=info msg="CreateContainer within sandbox \"36469c50d434b4437fc83a6d74c8b1e31eb7fb9d619e025a49624fe2ce7564b0\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"23bfa211c01ea6e7ad30d344e95bf8aa7807d5508a254ef4f31eb8d812c6b7e6\"" Sep 13 01:22:05.173073 containerd[1819]: time="2025-09-13T01:22:05.173058927Z" level=info msg="StartContainer for \"23bfa211c01ea6e7ad30d344e95bf8aa7807d5508a254ef4f31eb8d812c6b7e6\"" Sep 13 01:22:05.202416 systemd[1]: Started cri-containerd-23bfa211c01ea6e7ad30d344e95bf8aa7807d5508a254ef4f31eb8d812c6b7e6.scope - libcontainer container 23bfa211c01ea6e7ad30d344e95bf8aa7807d5508a254ef4f31eb8d812c6b7e6. Sep 13 01:22:05.222112 containerd[1819]: time="2025-09-13T01:22:05.222047722Z" level=info msg="StartContainer for \"23bfa211c01ea6e7ad30d344e95bf8aa7807d5508a254ef4f31eb8d812c6b7e6\" returns successfully" Sep 13 01:22:05.296318 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 13 01:22:05.296373 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 13 01:22:05.355501 containerd[1819]: time="2025-09-13T01:22:05.355437293Z" level=info msg="StopPodSandbox for \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\"" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.388 [INFO][4709] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.388 [INFO][4709] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" iface="eth0" netns="/var/run/netns/cni-e631714b-2868-84af-b885-cd861308a0fd" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.389 [INFO][4709] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" iface="eth0" netns="/var/run/netns/cni-e631714b-2868-84af-b885-cd861308a0fd" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.389 [INFO][4709] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" iface="eth0" netns="/var/run/netns/cni-e631714b-2868-84af-b885-cd861308a0fd" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.389 [INFO][4709] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.389 [INFO][4709] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.404 [INFO][4740] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.404 [INFO][4740] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.404 [INFO][4740] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.409 [WARNING][4740] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.409 [INFO][4740] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.410 [INFO][4740] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:05.413768 containerd[1819]: 2025-09-13 01:22:05.412 [INFO][4709] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:05.414133 containerd[1819]: time="2025-09-13T01:22:05.413831591Z" level=info msg="TearDown network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\" successfully" Sep 13 01:22:05.414133 containerd[1819]: time="2025-09-13T01:22:05.413855876Z" level=info msg="StopPodSandbox for \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\" returns successfully" Sep 13 01:22:05.463106 kubelet[3072]: I0913 01:22:05.463043 3072 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctrsl\" (UniqueName: \"kubernetes.io/projected/971bad55-1380-4bfd-8af5-2cbef17292ea-kube-api-access-ctrsl\") pod \"971bad55-1380-4bfd-8af5-2cbef17292ea\" (UID: \"971bad55-1380-4bfd-8af5-2cbef17292ea\") " Sep 13 01:22:05.463106 kubelet[3072]: I0913 01:22:05.463086 3072 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-ca-bundle\") pod \"971bad55-1380-4bfd-8af5-2cbef17292ea\" (UID: \"971bad55-1380-4bfd-8af5-2cbef17292ea\") " Sep 13 01:22:05.463106 kubelet[3072]: I0913 01:22:05.463108 3072 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-backend-key-pair\") pod \"971bad55-1380-4bfd-8af5-2cbef17292ea\" (UID: \"971bad55-1380-4bfd-8af5-2cbef17292ea\") " Sep 13 01:22:05.463514 kubelet[3072]: I0913 01:22:05.463402 3072 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "971bad55-1380-4bfd-8af5-2cbef17292ea" (UID: "971bad55-1380-4bfd-8af5-2cbef17292ea"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 13 01:22:05.464989 kubelet[3072]: I0913 01:22:05.464939 3072 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "971bad55-1380-4bfd-8af5-2cbef17292ea" (UID: "971bad55-1380-4bfd-8af5-2cbef17292ea"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 13 01:22:05.464989 kubelet[3072]: I0913 01:22:05.464956 3072 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/971bad55-1380-4bfd-8af5-2cbef17292ea-kube-api-access-ctrsl" (OuterVolumeSpecName: "kube-api-access-ctrsl") pod "971bad55-1380-4bfd-8af5-2cbef17292ea" (UID: "971bad55-1380-4bfd-8af5-2cbef17292ea"). InnerVolumeSpecName "kube-api-access-ctrsl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 01:22:05.564173 kubelet[3072]: I0913 01:22:05.564073 3072 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ctrsl\" (UniqueName: \"kubernetes.io/projected/971bad55-1380-4bfd-8af5-2cbef17292ea-kube-api-access-ctrsl\") on node \"ci-4081.3.5-n-7cd54b3887\" DevicePath \"\"" Sep 13 01:22:05.564173 kubelet[3072]: I0913 01:22:05.564143 3072 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-ca-bundle\") on node \"ci-4081.3.5-n-7cd54b3887\" DevicePath \"\"" Sep 13 01:22:05.564173 kubelet[3072]: I0913 01:22:05.564174 3072 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/971bad55-1380-4bfd-8af5-2cbef17292ea-whisker-backend-key-pair\") on node \"ci-4081.3.5-n-7cd54b3887\" DevicePath \"\"" Sep 13 01:22:05.733169 systemd[1]: Removed slice kubepods-besteffort-pod971bad55_1380_4bfd_8af5_2cbef17292ea.slice - libcontainer container kubepods-besteffort-pod971bad55_1380_4bfd_8af5_2cbef17292ea.slice. Sep 13 01:22:05.856855 kubelet[3072]: I0913 01:22:05.856772 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-lxfxq" podStartSLOduration=1.4797482450000001 podStartE2EDuration="16.856745036s" podCreationTimestamp="2025-09-13 01:21:49 +0000 UTC" firstStartedPulling="2025-09-13 01:21:49.787086165 +0000 UTC m=+16.112337615" lastFinishedPulling="2025-09-13 01:22:05.164082965 +0000 UTC m=+31.489334406" observedRunningTime="2025-09-13 01:22:05.855966185 +0000 UTC m=+32.181217686" watchObservedRunningTime="2025-09-13 01:22:05.856745036 +0000 UTC m=+32.181996504" Sep 13 01:22:05.898324 systemd[1]: Created slice kubepods-besteffort-pod7534749d_4b9d_4b37_89bf_4c51090be9a8.slice - libcontainer container kubepods-besteffort-pod7534749d_4b9d_4b37_89bf_4c51090be9a8.slice. Sep 13 01:22:05.967574 kubelet[3072]: I0913 01:22:05.967501 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7534749d-4b9d-4b37-89bf-4c51090be9a8-whisker-ca-bundle\") pod \"whisker-56d88b84f5-7zfgn\" (UID: \"7534749d-4b9d-4b37-89bf-4c51090be9a8\") " pod="calico-system/whisker-56d88b84f5-7zfgn" Sep 13 01:22:05.967828 kubelet[3072]: I0913 01:22:05.967670 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/7534749d-4b9d-4b37-89bf-4c51090be9a8-whisker-backend-key-pair\") pod \"whisker-56d88b84f5-7zfgn\" (UID: \"7534749d-4b9d-4b37-89bf-4c51090be9a8\") " pod="calico-system/whisker-56d88b84f5-7zfgn" Sep 13 01:22:05.967828 kubelet[3072]: I0913 01:22:05.967752 3072 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5nq5\" (UniqueName: \"kubernetes.io/projected/7534749d-4b9d-4b37-89bf-4c51090be9a8-kube-api-access-n5nq5\") pod \"whisker-56d88b84f5-7zfgn\" (UID: \"7534749d-4b9d-4b37-89bf-4c51090be9a8\") " pod="calico-system/whisker-56d88b84f5-7zfgn" Sep 13 01:22:06.152207 systemd[1]: run-netns-cni\x2de631714b\x2d2868\x2d84af\x2db885\x2dcd861308a0fd.mount: Deactivated successfully. Sep 13 01:22:06.152471 systemd[1]: var-lib-kubelet-pods-971bad55\x2d1380\x2d4bfd\x2d8af5\x2d2cbef17292ea-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dctrsl.mount: Deactivated successfully. Sep 13 01:22:06.152661 systemd[1]: var-lib-kubelet-pods-971bad55\x2d1380\x2d4bfd\x2d8af5\x2d2cbef17292ea-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 13 01:22:06.203279 containerd[1819]: time="2025-09-13T01:22:06.203122029Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-56d88b84f5-7zfgn,Uid:7534749d-4b9d-4b37-89bf-4c51090be9a8,Namespace:calico-system,Attempt:0,}" Sep 13 01:22:06.267229 systemd-networkd[1608]: calid3807015d89: Link UP Sep 13 01:22:06.267452 systemd-networkd[1608]: calid3807015d89: Gained carrier Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.219 [INFO][4803] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.226 [INFO][4803] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0 whisker-56d88b84f5- calico-system 7534749d-4b9d-4b37-89bf-4c51090be9a8 859 0 2025-09-13 01:22:05 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:56d88b84f5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 whisker-56d88b84f5-7zfgn eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calid3807015d89 [] [] }} ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.226 [INFO][4803] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.240 [INFO][4825] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" HandleID="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.240 [INFO][4825] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" HandleID="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f6c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"whisker-56d88b84f5-7zfgn", "timestamp":"2025-09-13 01:22:06.240386844 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.240 [INFO][4825] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.240 [INFO][4825] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.240 [INFO][4825] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.245 [INFO][4825] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.248 [INFO][4825] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.251 [INFO][4825] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.253 [INFO][4825] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.254 [INFO][4825] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.254 [INFO][4825] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.255 [INFO][4825] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9 Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.258 [INFO][4825] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.261 [INFO][4825] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.193/26] block=192.168.103.192/26 handle="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.261 [INFO][4825] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.193/26] handle="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.261 [INFO][4825] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:06.295913 containerd[1819]: 2025-09-13 01:22:06.261 [INFO][4825] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.193/26] IPv6=[] ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" HandleID="k8s-pod-network.364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" Sep 13 01:22:06.297885 containerd[1819]: 2025-09-13 01:22:06.262 [INFO][4803] cni-plugin/k8s.go 418: Populated endpoint ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0", GenerateName:"whisker-56d88b84f5-", Namespace:"calico-system", SelfLink:"", UID:"7534749d-4b9d-4b37-89bf-4c51090be9a8", ResourceVersion:"859", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"56d88b84f5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"whisker-56d88b84f5-7zfgn", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.103.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calid3807015d89", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:06.297885 containerd[1819]: 2025-09-13 01:22:06.262 [INFO][4803] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.193/32] ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" Sep 13 01:22:06.297885 containerd[1819]: 2025-09-13 01:22:06.262 [INFO][4803] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid3807015d89 ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" Sep 13 01:22:06.297885 containerd[1819]: 2025-09-13 01:22:06.267 [INFO][4803] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" Sep 13 01:22:06.297885 containerd[1819]: 2025-09-13 01:22:06.268 [INFO][4803] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0", GenerateName:"whisker-56d88b84f5-", Namespace:"calico-system", SelfLink:"", UID:"7534749d-4b9d-4b37-89bf-4c51090be9a8", ResourceVersion:"859", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 22, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"56d88b84f5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9", Pod:"whisker-56d88b84f5-7zfgn", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.103.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calid3807015d89", MAC:"3e:ee:58:24:30:34", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:06.297885 containerd[1819]: 2025-09-13 01:22:06.291 [INFO][4803] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9" Namespace="calico-system" Pod="whisker-56d88b84f5-7zfgn" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--56d88b84f5--7zfgn-eth0" Sep 13 01:22:06.308826 containerd[1819]: time="2025-09-13T01:22:06.308767349Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:06.308826 containerd[1819]: time="2025-09-13T01:22:06.308799316Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:06.308826 containerd[1819]: time="2025-09-13T01:22:06.308825460Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:06.308936 containerd[1819]: time="2025-09-13T01:22:06.308868096Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:06.333746 systemd[1]: Started cri-containerd-364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9.scope - libcontainer container 364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9. Sep 13 01:22:06.407323 containerd[1819]: time="2025-09-13T01:22:06.407225143Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-56d88b84f5-7zfgn,Uid:7534749d-4b9d-4b37-89bf-4c51090be9a8,Namespace:calico-system,Attempt:0,} returns sandbox id \"364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9\"" Sep 13 01:22:06.408275 containerd[1819]: time="2025-09-13T01:22:06.408227579Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 13 01:22:07.727378 kubelet[3072]: I0913 01:22:07.727299 3072 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="971bad55-1380-4bfd-8af5-2cbef17292ea" path="/var/lib/kubelet/pods/971bad55-1380-4bfd-8af5-2cbef17292ea/volumes" Sep 13 01:22:07.959657 containerd[1819]: time="2025-09-13T01:22:07.959632313Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:07.959874 containerd[1819]: time="2025-09-13T01:22:07.959830438Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4661291" Sep 13 01:22:07.960100 containerd[1819]: time="2025-09-13T01:22:07.960087100Z" level=info msg="ImageCreate event name:\"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:07.961664 containerd[1819]: time="2025-09-13T01:22:07.961652555Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:07.961951 containerd[1819]: time="2025-09-13T01:22:07.961938345Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"6153986\" in 1.553690164s" Sep 13 01:22:07.961973 containerd[1819]: time="2025-09-13T01:22:07.961955167Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\"" Sep 13 01:22:07.963036 containerd[1819]: time="2025-09-13T01:22:07.963025265Z" level=info msg="CreateContainer within sandbox \"364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 13 01:22:07.967306 containerd[1819]: time="2025-09-13T01:22:07.967290400Z" level=info msg="CreateContainer within sandbox \"364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"86c0b0f13eeb68d57473d6cbab68e697729ac60568a14aa1cbbb785211663b56\"" Sep 13 01:22:07.967568 containerd[1819]: time="2025-09-13T01:22:07.967498152Z" level=info msg="StartContainer for \"86c0b0f13eeb68d57473d6cbab68e697729ac60568a14aa1cbbb785211663b56\"" Sep 13 01:22:07.995506 systemd[1]: Started cri-containerd-86c0b0f13eeb68d57473d6cbab68e697729ac60568a14aa1cbbb785211663b56.scope - libcontainer container 86c0b0f13eeb68d57473d6cbab68e697729ac60568a14aa1cbbb785211663b56. Sep 13 01:22:08.021706 containerd[1819]: time="2025-09-13T01:22:08.021683313Z" level=info msg="StartContainer for \"86c0b0f13eeb68d57473d6cbab68e697729ac60568a14aa1cbbb785211663b56\" returns successfully" Sep 13 01:22:08.022295 containerd[1819]: time="2025-09-13T01:22:08.022279967Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 13 01:22:08.045461 systemd-networkd[1608]: calid3807015d89: Gained IPv6LL Sep 13 01:22:10.256913 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2558329780.mount: Deactivated successfully. Sep 13 01:22:10.262250 containerd[1819]: time="2025-09-13T01:22:10.262201631Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:10.262419 containerd[1819]: time="2025-09-13T01:22:10.262400233Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=33085545" Sep 13 01:22:10.262769 containerd[1819]: time="2025-09-13T01:22:10.262730259Z" level=info msg="ImageCreate event name:\"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:10.263884 containerd[1819]: time="2025-09-13T01:22:10.263845038Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:10.264295 containerd[1819]: time="2025-09-13T01:22:10.264264381Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"33085375\" in 2.241965204s" Sep 13 01:22:10.264295 containerd[1819]: time="2025-09-13T01:22:10.264282929Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\"" Sep 13 01:22:10.265253 containerd[1819]: time="2025-09-13T01:22:10.265233534Z" level=info msg="CreateContainer within sandbox \"364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 13 01:22:10.271121 containerd[1819]: time="2025-09-13T01:22:10.271081018Z" level=info msg="CreateContainer within sandbox \"364cd022f7d80be5f296d69f92a0b3caa929b3cf0c20d8b247cd10f76c4b3dc9\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"3e3f86f4bbaeb1d14fc6f2361b6b51edb4bb84f968e1685f9540cbfc17567ae9\"" Sep 13 01:22:10.271306 containerd[1819]: time="2025-09-13T01:22:10.271293556Z" level=info msg="StartContainer for \"3e3f86f4bbaeb1d14fc6f2361b6b51edb4bb84f968e1685f9540cbfc17567ae9\"" Sep 13 01:22:10.299532 systemd[1]: Started cri-containerd-3e3f86f4bbaeb1d14fc6f2361b6b51edb4bb84f968e1685f9540cbfc17567ae9.scope - libcontainer container 3e3f86f4bbaeb1d14fc6f2361b6b51edb4bb84f968e1685f9540cbfc17567ae9. Sep 13 01:22:10.325309 containerd[1819]: time="2025-09-13T01:22:10.325287536Z" level=info msg="StartContainer for \"3e3f86f4bbaeb1d14fc6f2361b6b51edb4bb84f968e1685f9540cbfc17567ae9\" returns successfully" Sep 13 01:22:10.868548 kubelet[3072]: I0913 01:22:10.868385 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-56d88b84f5-7zfgn" podStartSLOduration=2.011723478 podStartE2EDuration="5.868348027s" podCreationTimestamp="2025-09-13 01:22:05 +0000 UTC" firstStartedPulling="2025-09-13 01:22:06.408052169 +0000 UTC m=+32.733303617" lastFinishedPulling="2025-09-13 01:22:10.264676722 +0000 UTC m=+36.589928166" observedRunningTime="2025-09-13 01:22:10.86813751 +0000 UTC m=+37.193389023" watchObservedRunningTime="2025-09-13 01:22:10.868348027 +0000 UTC m=+37.193599520" Sep 13 01:22:11.721453 containerd[1819]: time="2025-09-13T01:22:11.721412905Z" level=info msg="StopPodSandbox for \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\"" Sep 13 01:22:11.721669 containerd[1819]: time="2025-09-13T01:22:11.721412105Z" level=info msg="StopPodSandbox for \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\"" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5441] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5441] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" iface="eth0" netns="/var/run/netns/cni-c7df442e-43d6-d503-8be9-2c98d72b8700" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5441] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" iface="eth0" netns="/var/run/netns/cni-c7df442e-43d6-d503-8be9-2c98d72b8700" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5441] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" iface="eth0" netns="/var/run/netns/cni-c7df442e-43d6-d503-8be9-2c98d72b8700" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5441] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5441] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.753 [INFO][5471] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.753 [INFO][5471] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.753 [INFO][5471] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.757 [WARNING][5471] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.757 [INFO][5471] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.758 [INFO][5471] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:11.759565 containerd[1819]: 2025-09-13 01:22:11.758 [INFO][5441] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:11.759862 containerd[1819]: time="2025-09-13T01:22:11.759648716Z" level=info msg="TearDown network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\" successfully" Sep 13 01:22:11.759862 containerd[1819]: time="2025-09-13T01:22:11.759669487Z" level=info msg="StopPodSandbox for \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\" returns successfully" Sep 13 01:22:11.760043 containerd[1819]: time="2025-09-13T01:22:11.760029510Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5hgc5,Uid:465526ae-5203-429b-8157-1675929b92b0,Namespace:kube-system,Attempt:1,}" Sep 13 01:22:11.761201 systemd[1]: run-netns-cni\x2dc7df442e\x2d43d6\x2dd503\x2d8be9\x2d2c98d72b8700.mount: Deactivated successfully. Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5440] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5440] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" iface="eth0" netns="/var/run/netns/cni-5cd8dd86-3226-5973-7e8b-13c87d991a73" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5440] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" iface="eth0" netns="/var/run/netns/cni-5cd8dd86-3226-5973-7e8b-13c87d991a73" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5440] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" iface="eth0" netns="/var/run/netns/cni-5cd8dd86-3226-5973-7e8b-13c87d991a73" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5440] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.743 [INFO][5440] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.754 [INFO][5469] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.754 [INFO][5469] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.758 [INFO][5469] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.761 [WARNING][5469] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.761 [INFO][5469] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.762 [INFO][5469] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:11.763653 containerd[1819]: 2025-09-13 01:22:11.763 [INFO][5440] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:11.763904 containerd[1819]: time="2025-09-13T01:22:11.763716096Z" level=info msg="TearDown network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\" successfully" Sep 13 01:22:11.763904 containerd[1819]: time="2025-09-13T01:22:11.763730654Z" level=info msg="StopPodSandbox for \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\" returns successfully" Sep 13 01:22:11.764040 containerd[1819]: time="2025-09-13T01:22:11.764027850Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-sfpqb,Uid:3999b298-1754-45b1-95dc-2f560b55d1e4,Namespace:calico-system,Attempt:1,}" Sep 13 01:22:11.767495 systemd[1]: run-netns-cni\x2d5cd8dd86\x2d3226\x2d5973\x2d7e8b\x2d13c87d991a73.mount: Deactivated successfully. Sep 13 01:22:11.809128 systemd-networkd[1608]: calia3c243e27ae: Link UP Sep 13 01:22:11.809269 systemd-networkd[1608]: calia3c243e27ae: Gained carrier Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.774 [INFO][5502] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.780 [INFO][5502] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0 coredns-668d6bf9bc- kube-system 465526ae-5203-429b-8157-1675929b92b0 891 0 2025-09-13 01:21:39 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 coredns-668d6bf9bc-5hgc5 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calia3c243e27ae [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.780 [INFO][5502] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.792 [INFO][5550] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" HandleID="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.792 [INFO][5550] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" HandleID="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00069e9d0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"coredns-668d6bf9bc-5hgc5", "timestamp":"2025-09-13 01:22:11.792574894 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.792 [INFO][5550] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.792 [INFO][5550] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.792 [INFO][5550] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.796 [INFO][5550] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.799 [INFO][5550] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.801 [INFO][5550] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.801 [INFO][5550] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.802 [INFO][5550] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.802 [INFO][5550] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.803 [INFO][5550] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.805 [INFO][5550] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.807 [INFO][5550] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.194/26] block=192.168.103.192/26 handle="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.807 [INFO][5550] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.194/26] handle="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.807 [INFO][5550] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:11.814998 containerd[1819]: 2025-09-13 01:22:11.807 [INFO][5550] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.194/26] IPv6=[] ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" HandleID="k8s-pod-network.969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.815415 containerd[1819]: 2025-09-13 01:22:11.808 [INFO][5502] cni-plugin/k8s.go 418: Populated endpoint ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"465526ae-5203-429b-8157-1675929b92b0", ResourceVersion:"891", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"coredns-668d6bf9bc-5hgc5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia3c243e27ae", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:11.815415 containerd[1819]: 2025-09-13 01:22:11.808 [INFO][5502] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.194/32] ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.815415 containerd[1819]: 2025-09-13 01:22:11.808 [INFO][5502] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia3c243e27ae ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.815415 containerd[1819]: 2025-09-13 01:22:11.809 [INFO][5502] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.815415 containerd[1819]: 2025-09-13 01:22:11.809 [INFO][5502] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"465526ae-5203-429b-8157-1675929b92b0", ResourceVersion:"891", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b", Pod:"coredns-668d6bf9bc-5hgc5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia3c243e27ae", MAC:"4e:67:8e:5f:62:ec", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:11.815415 containerd[1819]: 2025-09-13 01:22:11.813 [INFO][5502] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b" Namespace="kube-system" Pod="coredns-668d6bf9bc-5hgc5" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:11.823030 containerd[1819]: time="2025-09-13T01:22:11.822952574Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:11.823030 containerd[1819]: time="2025-09-13T01:22:11.822990879Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:11.823183 containerd[1819]: time="2025-09-13T01:22:11.823005512Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:11.823228 containerd[1819]: time="2025-09-13T01:22:11.823218385Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:11.846702 systemd[1]: Started cri-containerd-969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b.scope - libcontainer container 969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b. Sep 13 01:22:11.892797 containerd[1819]: time="2025-09-13T01:22:11.892771290Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5hgc5,Uid:465526ae-5203-429b-8157-1675929b92b0,Namespace:kube-system,Attempt:1,} returns sandbox id \"969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b\"" Sep 13 01:22:11.894033 containerd[1819]: time="2025-09-13T01:22:11.894016474Z" level=info msg="CreateContainer within sandbox \"969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 01:22:11.898447 containerd[1819]: time="2025-09-13T01:22:11.898395705Z" level=info msg="CreateContainer within sandbox \"969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4d750be08e49b9ef13ddb06702ca2d2298bd48921a45aa13a40a4d58e66a38ff\"" Sep 13 01:22:11.898639 containerd[1819]: time="2025-09-13T01:22:11.898623378Z" level=info msg="StartContainer for \"4d750be08e49b9ef13ddb06702ca2d2298bd48921a45aa13a40a4d58e66a38ff\"" Sep 13 01:22:11.911547 systemd-networkd[1608]: calibfbc197f5b1: Link UP Sep 13 01:22:11.911999 systemd-networkd[1608]: calibfbc197f5b1: Gained carrier Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.778 [INFO][5517] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.783 [INFO][5517] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0 goldmane-54d579b49d- calico-system 3999b298-1754-45b1-95dc-2f560b55d1e4 890 0 2025-09-13 01:21:48 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 goldmane-54d579b49d-sfpqb eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calibfbc197f5b1 [] [] }} ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.783 [INFO][5517] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.794 [INFO][5556] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" HandleID="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.794 [INFO][5556] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" HandleID="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000345a80), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"goldmane-54d579b49d-sfpqb", "timestamp":"2025-09-13 01:22:11.794694837 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.794 [INFO][5556] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.807 [INFO][5556] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.807 [INFO][5556] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.897 [INFO][5556] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.899 [INFO][5556] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.901 [INFO][5556] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.903 [INFO][5556] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.904 [INFO][5556] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.904 [INFO][5556] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.905 [INFO][5556] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.907 [INFO][5556] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.909 [INFO][5556] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.195/26] block=192.168.103.192/26 handle="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.909 [INFO][5556] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.195/26] handle="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.909 [INFO][5556] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:11.917745 containerd[1819]: 2025-09-13 01:22:11.909 [INFO][5556] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.195/26] IPv6=[] ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" HandleID="k8s-pod-network.a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.918154 containerd[1819]: 2025-09-13 01:22:11.910 [INFO][5517] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"3999b298-1754-45b1-95dc-2f560b55d1e4", ResourceVersion:"890", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"goldmane-54d579b49d-sfpqb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calibfbc197f5b1", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:11.918154 containerd[1819]: 2025-09-13 01:22:11.910 [INFO][5517] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.195/32] ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.918154 containerd[1819]: 2025-09-13 01:22:11.910 [INFO][5517] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibfbc197f5b1 ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.918154 containerd[1819]: 2025-09-13 01:22:11.912 [INFO][5517] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.918154 containerd[1819]: 2025-09-13 01:22:11.912 [INFO][5517] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"3999b298-1754-45b1-95dc-2f560b55d1e4", ResourceVersion:"890", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d", Pod:"goldmane-54d579b49d-sfpqb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calibfbc197f5b1", MAC:"e2:27:43:04:e1:0f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:11.918154 containerd[1819]: 2025-09-13 01:22:11.916 [INFO][5517] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d" Namespace="calico-system" Pod="goldmane-54d579b49d-sfpqb" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:11.921432 systemd[1]: Started cri-containerd-4d750be08e49b9ef13ddb06702ca2d2298bd48921a45aa13a40a4d58e66a38ff.scope - libcontainer container 4d750be08e49b9ef13ddb06702ca2d2298bd48921a45aa13a40a4d58e66a38ff. Sep 13 01:22:11.926430 containerd[1819]: time="2025-09-13T01:22:11.926389854Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:11.926430 containerd[1819]: time="2025-09-13T01:22:11.926420687Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:11.926430 containerd[1819]: time="2025-09-13T01:22:11.926427645Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:11.926535 containerd[1819]: time="2025-09-13T01:22:11.926502682Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:11.932790 containerd[1819]: time="2025-09-13T01:22:11.932771134Z" level=info msg="StartContainer for \"4d750be08e49b9ef13ddb06702ca2d2298bd48921a45aa13a40a4d58e66a38ff\" returns successfully" Sep 13 01:22:11.952811 systemd[1]: Started cri-containerd-a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d.scope - libcontainer container a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d. Sep 13 01:22:12.016531 containerd[1819]: time="2025-09-13T01:22:12.016478479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-sfpqb,Uid:3999b298-1754-45b1-95dc-2f560b55d1e4,Namespace:calico-system,Attempt:1,} returns sandbox id \"a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d\"" Sep 13 01:22:12.017335 containerd[1819]: time="2025-09-13T01:22:12.017288234Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 13 01:22:12.880175 kubelet[3072]: I0913 01:22:12.880057 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-5hgc5" podStartSLOduration=33.880021687 podStartE2EDuration="33.880021687s" podCreationTimestamp="2025-09-13 01:21:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:22:12.879527592 +0000 UTC m=+39.204779105" watchObservedRunningTime="2025-09-13 01:22:12.880021687 +0000 UTC m=+39.205273189" Sep 13 01:22:13.549664 systemd-networkd[1608]: calibfbc197f5b1: Gained IPv6LL Sep 13 01:22:13.678451 systemd-networkd[1608]: calia3c243e27ae: Gained IPv6LL Sep 13 01:22:13.723398 containerd[1819]: time="2025-09-13T01:22:13.723324789Z" level=info msg="StopPodSandbox for \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\"" Sep 13 01:22:13.724603 containerd[1819]: time="2025-09-13T01:22:13.723345262Z" level=info msg="StopPodSandbox for \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\"" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5800] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5800] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" iface="eth0" netns="/var/run/netns/cni-230591ad-5a41-16ae-3feb-75df7048cf8c" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5800] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" iface="eth0" netns="/var/run/netns/cni-230591ad-5a41-16ae-3feb-75df7048cf8c" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5800] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" iface="eth0" netns="/var/run/netns/cni-230591ad-5a41-16ae-3feb-75df7048cf8c" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5800] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.779 [INFO][5800] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.789 [INFO][5875] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.789 [INFO][5875] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.789 [INFO][5875] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.792 [WARNING][5875] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.792 [INFO][5875] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.793 [INFO][5875] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:13.794748 containerd[1819]: 2025-09-13 01:22:13.794 [INFO][5800] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:13.795040 containerd[1819]: time="2025-09-13T01:22:13.794863687Z" level=info msg="TearDown network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\" successfully" Sep 13 01:22:13.795040 containerd[1819]: time="2025-09-13T01:22:13.794905451Z" level=info msg="StopPodSandbox for \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\" returns successfully" Sep 13 01:22:13.797458 systemd[1]: run-netns-cni\x2d230591ad\x2d5a41\x2d16ae\x2d3feb\x2d75df7048cf8c.mount: Deactivated successfully. Sep 13 01:22:13.797831 containerd[1819]: time="2025-09-13T01:22:13.797789615Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-rqw44,Uid:1e417130-1b3b-4716-84aa-d920f0d71e2e,Namespace:calico-apiserver,Attempt:1,}" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5799] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5799] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" iface="eth0" netns="/var/run/netns/cni-cde7912d-80cf-bc0e-6a61-c7a27ad374df" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5799] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" iface="eth0" netns="/var/run/netns/cni-cde7912d-80cf-bc0e-6a61-c7a27ad374df" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5799] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" iface="eth0" netns="/var/run/netns/cni-cde7912d-80cf-bc0e-6a61-c7a27ad374df" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5799] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.778 [INFO][5799] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.789 [INFO][5873] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.789 [INFO][5873] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.793 [INFO][5873] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.797 [WARNING][5873] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.797 [INFO][5873] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.798 [INFO][5873] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:13.799822 containerd[1819]: 2025-09-13 01:22:13.799 [INFO][5799] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:13.800115 containerd[1819]: time="2025-09-13T01:22:13.799862914Z" level=info msg="TearDown network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\" successfully" Sep 13 01:22:13.800115 containerd[1819]: time="2025-09-13T01:22:13.799875502Z" level=info msg="StopPodSandbox for \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\" returns successfully" Sep 13 01:22:13.800152 containerd[1819]: time="2025-09-13T01:22:13.800138337Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-689db9d4f-r55km,Uid:198f178f-c689-48cb-b5e4-012f47ac61f3,Namespace:calico-system,Attempt:1,}" Sep 13 01:22:13.801645 systemd[1]: run-netns-cni\x2dcde7912d\x2d80cf\x2dbc0e\x2d6a61\x2dc7a27ad374df.mount: Deactivated successfully. Sep 13 01:22:13.856405 systemd-networkd[1608]: calia435cbd0940: Link UP Sep 13 01:22:13.856580 systemd-networkd[1608]: calia435cbd0940: Gained carrier Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.817 [INFO][5915] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.823 [INFO][5915] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0 calico-kube-controllers-689db9d4f- calico-system 198f178f-c689-48cb-b5e4-012f47ac61f3 919 0 2025-09-13 01:21:49 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:689db9d4f projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 calico-kube-controllers-689db9d4f-r55km eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calia435cbd0940 [] [] }} ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.823 [INFO][5915] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5950] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" HandleID="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5950] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" HandleID="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f750), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"calico-kube-controllers-689db9d4f-r55km", "timestamp":"2025-09-13 01:22:13.835346157 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5950] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5950] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5950] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.840 [INFO][5950] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.843 [INFO][5950] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.846 [INFO][5950] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.847 [INFO][5950] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.848 [INFO][5950] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.848 [INFO][5950] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.849 [INFO][5950] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03 Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.851 [INFO][5950] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.854 [INFO][5950] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.196/26] block=192.168.103.192/26 handle="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.854 [INFO][5950] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.196/26] handle="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.854 [INFO][5950] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:13.862143 containerd[1819]: 2025-09-13 01:22:13.854 [INFO][5950] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.196/26] IPv6=[] ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" HandleID="k8s-pod-network.543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.862682 containerd[1819]: 2025-09-13 01:22:13.855 [INFO][5915] cni-plugin/k8s.go 418: Populated endpoint ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0", GenerateName:"calico-kube-controllers-689db9d4f-", Namespace:"calico-system", SelfLink:"", UID:"198f178f-c689-48cb-b5e4-012f47ac61f3", ResourceVersion:"919", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"689db9d4f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"calico-kube-controllers-689db9d4f-r55km", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calia435cbd0940", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:13.862682 containerd[1819]: 2025-09-13 01:22:13.855 [INFO][5915] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.196/32] ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.862682 containerd[1819]: 2025-09-13 01:22:13.855 [INFO][5915] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia435cbd0940 ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.862682 containerd[1819]: 2025-09-13 01:22:13.856 [INFO][5915] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.862682 containerd[1819]: 2025-09-13 01:22:13.856 [INFO][5915] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0", GenerateName:"calico-kube-controllers-689db9d4f-", Namespace:"calico-system", SelfLink:"", UID:"198f178f-c689-48cb-b5e4-012f47ac61f3", ResourceVersion:"919", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"689db9d4f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03", Pod:"calico-kube-controllers-689db9d4f-r55km", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calia435cbd0940", MAC:"0e:cc:2e:48:14:b3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:13.862682 containerd[1819]: 2025-09-13 01:22:13.861 [INFO][5915] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03" Namespace="calico-system" Pod="calico-kube-controllers-689db9d4f-r55km" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:13.871951 containerd[1819]: time="2025-09-13T01:22:13.869985997Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:13.871951 containerd[1819]: time="2025-09-13T01:22:13.870207190Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:13.871951 containerd[1819]: time="2025-09-13T01:22:13.870219852Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:13.871951 containerd[1819]: time="2025-09-13T01:22:13.870269321Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:13.896803 systemd[1]: Started cri-containerd-543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03.scope - libcontainer container 543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03. Sep 13 01:22:13.956157 containerd[1819]: time="2025-09-13T01:22:13.956109710Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-689db9d4f-r55km,Uid:198f178f-c689-48cb-b5e4-012f47ac61f3,Namespace:calico-system,Attempt:1,} returns sandbox id \"543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03\"" Sep 13 01:22:13.976073 systemd-networkd[1608]: calif454ca88c93: Link UP Sep 13 01:22:13.976239 systemd-networkd[1608]: calif454ca88c93: Gained carrier Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.816 [INFO][5905] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.822 [INFO][5905] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0 calico-apiserver-db75486c4- calico-apiserver 1e417130-1b3b-4716-84aa-d920f0d71e2e 918 0 2025-09-13 01:21:47 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:db75486c4 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 calico-apiserver-db75486c4-rqw44 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calif454ca88c93 [] [] }} ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.822 [INFO][5905] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5948] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" HandleID="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5948] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" HandleID="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f520), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"calico-apiserver-db75486c4-rqw44", "timestamp":"2025-09-13 01:22:13.835346174 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.835 [INFO][5948] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.854 [INFO][5948] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.854 [INFO][5948] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.941 [INFO][5948] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.945 [INFO][5948] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.949 [INFO][5948] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.950 [INFO][5948] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.952 [INFO][5948] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.952 [INFO][5948] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.953 [INFO][5948] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1 Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.970 [INFO][5948] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.973 [INFO][5948] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.197/26] block=192.168.103.192/26 handle="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.974 [INFO][5948] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.197/26] handle="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.974 [INFO][5948] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:13.981571 containerd[1819]: 2025-09-13 01:22:13.974 [INFO][5948] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.197/26] IPv6=[] ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" HandleID="k8s-pod-network.0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.982025 containerd[1819]: 2025-09-13 01:22:13.975 [INFO][5905] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1e417130-1b3b-4716-84aa-d920f0d71e2e", ResourceVersion:"918", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"calico-apiserver-db75486c4-rqw44", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif454ca88c93", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:13.982025 containerd[1819]: 2025-09-13 01:22:13.975 [INFO][5905] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.197/32] ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.982025 containerd[1819]: 2025-09-13 01:22:13.975 [INFO][5905] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif454ca88c93 ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.982025 containerd[1819]: 2025-09-13 01:22:13.976 [INFO][5905] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.982025 containerd[1819]: 2025-09-13 01:22:13.976 [INFO][5905] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1e417130-1b3b-4716-84aa-d920f0d71e2e", ResourceVersion:"918", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1", Pod:"calico-apiserver-db75486c4-rqw44", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif454ca88c93", MAC:"3a:20:ee:82:3b:73", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:13.982025 containerd[1819]: 2025-09-13 01:22:13.980 [INFO][5905] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-rqw44" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:13.989958 containerd[1819]: time="2025-09-13T01:22:13.989755781Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:13.989958 containerd[1819]: time="2025-09-13T01:22:13.989949906Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:13.990053 containerd[1819]: time="2025-09-13T01:22:13.989958178Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:13.990053 containerd[1819]: time="2025-09-13T01:22:13.990001231Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:14.010390 systemd[1]: Started cri-containerd-0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1.scope - libcontainer container 0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1. Sep 13 01:22:14.037401 containerd[1819]: time="2025-09-13T01:22:14.037351786Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-rqw44,Uid:1e417130-1b3b-4716-84aa-d920f0d71e2e,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1\"" Sep 13 01:22:14.589711 containerd[1819]: time="2025-09-13T01:22:14.589686329Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:14.589928 containerd[1819]: time="2025-09-13T01:22:14.589904595Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=66357526" Sep 13 01:22:14.590216 containerd[1819]: time="2025-09-13T01:22:14.590205108Z" level=info msg="ImageCreate event name:\"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:14.591420 containerd[1819]: time="2025-09-13T01:22:14.591405829Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:14.591962 containerd[1819]: time="2025-09-13T01:22:14.591909020Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"66357372\" in 2.574602153s" Sep 13 01:22:14.591962 containerd[1819]: time="2025-09-13T01:22:14.591924661Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\"" Sep 13 01:22:14.592464 containerd[1819]: time="2025-09-13T01:22:14.592452953Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 13 01:22:14.592928 containerd[1819]: time="2025-09-13T01:22:14.592918279Z" level=info msg="CreateContainer within sandbox \"a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 13 01:22:14.597249 containerd[1819]: time="2025-09-13T01:22:14.597203753Z" level=info msg="CreateContainer within sandbox \"a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"decb93d2ad1ebd13225b8daff4698537418b39bbf002a26e0fafc11d66e2dfe8\"" Sep 13 01:22:14.597481 containerd[1819]: time="2025-09-13T01:22:14.597439983Z" level=info msg="StartContainer for \"decb93d2ad1ebd13225b8daff4698537418b39bbf002a26e0fafc11d66e2dfe8\"" Sep 13 01:22:14.627530 systemd[1]: Started cri-containerd-decb93d2ad1ebd13225b8daff4698537418b39bbf002a26e0fafc11d66e2dfe8.scope - libcontainer container decb93d2ad1ebd13225b8daff4698537418b39bbf002a26e0fafc11d66e2dfe8. Sep 13 01:22:14.657030 containerd[1819]: time="2025-09-13T01:22:14.656979730Z" level=info msg="StartContainer for \"decb93d2ad1ebd13225b8daff4698537418b39bbf002a26e0fafc11d66e2dfe8\" returns successfully" Sep 13 01:22:14.721318 containerd[1819]: time="2025-09-13T01:22:14.721280979Z" level=info msg="StopPodSandbox for \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\"" Sep 13 01:22:14.721318 containerd[1819]: time="2025-09-13T01:22:14.721297828Z" level=info msg="StopPodSandbox for \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\"" Sep 13 01:22:14.721517 containerd[1819]: time="2025-09-13T01:22:14.721372551Z" level=info msg="StopPodSandbox for \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\"" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6173] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6173] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" iface="eth0" netns="/var/run/netns/cni-fd66a618-3e21-c92a-71fb-353578de10fd" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6173] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" iface="eth0" netns="/var/run/netns/cni-fd66a618-3e21-c92a-71fb-353578de10fd" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6173] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" iface="eth0" netns="/var/run/netns/cni-fd66a618-3e21-c92a-71fb-353578de10fd" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6173] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6173] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.781 [INFO][6208] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.781 [INFO][6208] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.781 [INFO][6208] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.786 [WARNING][6208] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.786 [INFO][6208] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.787 [INFO][6208] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:14.789392 containerd[1819]: 2025-09-13 01:22:14.788 [INFO][6173] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:14.789852 containerd[1819]: time="2025-09-13T01:22:14.789831065Z" level=info msg="TearDown network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\" successfully" Sep 13 01:22:14.789852 containerd[1819]: time="2025-09-13T01:22:14.789851015Z" level=info msg="StopPodSandbox for \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\" returns successfully" Sep 13 01:22:14.790353 containerd[1819]: time="2025-09-13T01:22:14.790336047Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-zkzzg,Uid:0db74307-721a-4439-bd8d-982b96aa7db8,Namespace:kube-system,Attempt:1,}" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6174] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6174] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" iface="eth0" netns="/var/run/netns/cni-4c9cd8b0-ff3b-a8d5-b382-787c7deaa4a9" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6174] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" iface="eth0" netns="/var/run/netns/cni-4c9cd8b0-ff3b-a8d5-b382-787c7deaa4a9" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6174] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" iface="eth0" netns="/var/run/netns/cni-4c9cd8b0-ff3b-a8d5-b382-787c7deaa4a9" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6174] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.769 [INFO][6174] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.782 [INFO][6207] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.782 [INFO][6207] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.787 [INFO][6207] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.792 [WARNING][6207] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.792 [INFO][6207] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.793 [INFO][6207] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:14.799084 containerd[1819]: 2025-09-13 01:22:14.797 [INFO][6174] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:14.799373 containerd[1819]: time="2025-09-13T01:22:14.799176102Z" level=info msg="TearDown network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\" successfully" Sep 13 01:22:14.799373 containerd[1819]: time="2025-09-13T01:22:14.799200092Z" level=info msg="StopPodSandbox for \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\" returns successfully" Sep 13 01:22:14.799645 containerd[1819]: time="2025-09-13T01:22:14.799630630Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-jqdjm,Uid:f4d72f52-2d93-4ed1-9fa8-3baa96684daf,Namespace:calico-system,Attempt:1,}" Sep 13 01:22:14.801469 systemd[1]: run-netns-cni\x2dfd66a618\x2d3e21\x2dc92a\x2d71fb\x2d353578de10fd.mount: Deactivated successfully. Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.770 [INFO][6175] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.771 [INFO][6175] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" iface="eth0" netns="/var/run/netns/cni-58c41810-ea02-90bb-efbd-d0db5891e644" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.771 [INFO][6175] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" iface="eth0" netns="/var/run/netns/cni-58c41810-ea02-90bb-efbd-d0db5891e644" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.771 [INFO][6175] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" iface="eth0" netns="/var/run/netns/cni-58c41810-ea02-90bb-efbd-d0db5891e644" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.771 [INFO][6175] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.771 [INFO][6175] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.782 [INFO][6218] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.782 [INFO][6218] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.793 [INFO][6218] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.797 [WARNING][6218] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.797 [INFO][6218] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.798 [INFO][6218] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:14.804292 containerd[1819]: 2025-09-13 01:22:14.802 [INFO][6175] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:14.804628 containerd[1819]: time="2025-09-13T01:22:14.804405278Z" level=info msg="TearDown network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\" successfully" Sep 13 01:22:14.804628 containerd[1819]: time="2025-09-13T01:22:14.804427907Z" level=info msg="StopPodSandbox for \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\" returns successfully" Sep 13 01:22:14.804959 containerd[1819]: time="2025-09-13T01:22:14.804939492Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-sm7wq,Uid:878b3bf4-9638-46e7-9a71-02b8d2591f7f,Namespace:calico-apiserver,Attempt:1,}" Sep 13 01:22:14.805772 systemd[1]: run-netns-cni\x2d4c9cd8b0\x2dff3b\x2da8d5\x2db382\x2d787c7deaa4a9.mount: Deactivated successfully. Sep 13 01:22:14.809818 systemd[1]: run-netns-cni\x2d58c41810\x2dea02\x2d90bb\x2defbd\x2dd0db5891e644.mount: Deactivated successfully. Sep 13 01:22:14.853686 systemd-networkd[1608]: cali3d80864d2e9: Link UP Sep 13 01:22:14.853829 systemd-networkd[1608]: cali3d80864d2e9: Gained carrier Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.811 [INFO][6288] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.818 [INFO][6288] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0 coredns-668d6bf9bc- kube-system 0db74307-721a-4439-bd8d-982b96aa7db8 936 0 2025-09-13 01:21:39 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 coredns-668d6bf9bc-zkzzg eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali3d80864d2e9 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.818 [INFO][6288] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.832 [INFO][6347] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" HandleID="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.832 [INFO][6347] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" HandleID="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002e6600), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"coredns-668d6bf9bc-zkzzg", "timestamp":"2025-09-13 01:22:14.832674269 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.832 [INFO][6347] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.832 [INFO][6347] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.832 [INFO][6347] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.836 [INFO][6347] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.839 [INFO][6347] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.842 [INFO][6347] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.844 [INFO][6347] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.845 [INFO][6347] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.845 [INFO][6347] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.846 [INFO][6347] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.848 [INFO][6347] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.851 [INFO][6347] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.198/26] block=192.168.103.192/26 handle="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.851 [INFO][6347] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.198/26] handle="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.851 [INFO][6347] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:14.860131 containerd[1819]: 2025-09-13 01:22:14.851 [INFO][6347] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.198/26] IPv6=[] ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" HandleID="k8s-pod-network.8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.860643 containerd[1819]: 2025-09-13 01:22:14.852 [INFO][6288] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0db74307-721a-4439-bd8d-982b96aa7db8", ResourceVersion:"936", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"coredns-668d6bf9bc-zkzzg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3d80864d2e9", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:14.860643 containerd[1819]: 2025-09-13 01:22:14.852 [INFO][6288] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.198/32] ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.860643 containerd[1819]: 2025-09-13 01:22:14.852 [INFO][6288] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3d80864d2e9 ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.860643 containerd[1819]: 2025-09-13 01:22:14.853 [INFO][6288] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.860643 containerd[1819]: 2025-09-13 01:22:14.853 [INFO][6288] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0db74307-721a-4439-bd8d-982b96aa7db8", ResourceVersion:"936", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc", Pod:"coredns-668d6bf9bc-zkzzg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3d80864d2e9", MAC:"fa:ab:d5:c1:53:71", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:14.860643 containerd[1819]: 2025-09-13 01:22:14.859 [INFO][6288] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc" Namespace="kube-system" Pod="coredns-668d6bf9bc-zkzzg" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:14.868060 kubelet[3072]: I0913 01:22:14.868020 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-sfpqb" podStartSLOduration=24.292758603 podStartE2EDuration="26.868005372s" podCreationTimestamp="2025-09-13 01:21:48 +0000 UTC" firstStartedPulling="2025-09-13 01:22:12.017135534 +0000 UTC m=+38.342386980" lastFinishedPulling="2025-09-13 01:22:14.592382306 +0000 UTC m=+40.917633749" observedRunningTime="2025-09-13 01:22:14.867701465 +0000 UTC m=+41.192952909" watchObservedRunningTime="2025-09-13 01:22:14.868005372 +0000 UTC m=+41.193256813" Sep 13 01:22:14.868750 containerd[1819]: time="2025-09-13T01:22:14.868695074Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:14.868799 containerd[1819]: time="2025-09-13T01:22:14.868750627Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:14.868957 containerd[1819]: time="2025-09-13T01:22:14.868945332Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:14.869019 containerd[1819]: time="2025-09-13T01:22:14.869007845Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:14.885363 systemd[1]: Started cri-containerd-8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc.scope - libcontainer container 8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc. Sep 13 01:22:14.893355 systemd-networkd[1608]: calia435cbd0940: Gained IPv6LL Sep 13 01:22:14.910351 containerd[1819]: time="2025-09-13T01:22:14.910330509Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-zkzzg,Uid:0db74307-721a-4439-bd8d-982b96aa7db8,Namespace:kube-system,Attempt:1,} returns sandbox id \"8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc\"" Sep 13 01:22:14.911551 containerd[1819]: time="2025-09-13T01:22:14.911537453Z" level=info msg="CreateContainer within sandbox \"8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 01:22:14.932835 containerd[1819]: time="2025-09-13T01:22:14.932782243Z" level=info msg="CreateContainer within sandbox \"8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d3393ed809f7c32981f7b611f071196fb3620b2754431f5c81e9ce4511e8f670\"" Sep 13 01:22:14.933035 containerd[1819]: time="2025-09-13T01:22:14.933022075Z" level=info msg="StartContainer for \"d3393ed809f7c32981f7b611f071196fb3620b2754431f5c81e9ce4511e8f670\"" Sep 13 01:22:14.948434 systemd[1]: Started cri-containerd-d3393ed809f7c32981f7b611f071196fb3620b2754431f5c81e9ce4511e8f670.scope - libcontainer container d3393ed809f7c32981f7b611f071196fb3620b2754431f5c81e9ce4511e8f670. Sep 13 01:22:14.951690 systemd-networkd[1608]: calief24478d807: Link UP Sep 13 01:22:14.951856 systemd-networkd[1608]: calief24478d807: Gained carrier Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.817 [INFO][6300] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.823 [INFO][6300] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0 csi-node-driver- calico-system f4d72f52-2d93-4ed1-9fa8-3baa96684daf 935 0 2025-09-13 01:21:49 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 csi-node-driver-jqdjm eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calief24478d807 [] [] }} ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.823 [INFO][6300] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.836 [INFO][6354] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" HandleID="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.836 [INFO][6354] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" HandleID="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001395a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"csi-node-driver-jqdjm", "timestamp":"2025-09-13 01:22:14.836286517 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.836 [INFO][6354] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.851 [INFO][6354] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.851 [INFO][6354] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.936 [INFO][6354] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.940 [INFO][6354] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.942 [INFO][6354] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.943 [INFO][6354] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.944 [INFO][6354] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.944 [INFO][6354] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.945 [INFO][6354] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8 Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.947 [INFO][6354] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6354] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.199/26] block=192.168.103.192/26 handle="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6354] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.199/26] handle="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6354] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:14.957269 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6354] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.199/26] IPv6=[] ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" HandleID="k8s-pod-network.48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.957950 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6300] cni-plugin/k8s.go 418: Populated endpoint ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f4d72f52-2d93-4ed1-9fa8-3baa96684daf", ResourceVersion:"935", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"csi-node-driver-jqdjm", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief24478d807", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:14.957950 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6300] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.199/32] ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.957950 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6300] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calief24478d807 ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.957950 containerd[1819]: 2025-09-13 01:22:14.951 [INFO][6300] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.957950 containerd[1819]: 2025-09-13 01:22:14.952 [INFO][6300] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f4d72f52-2d93-4ed1-9fa8-3baa96684daf", ResourceVersion:"935", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8", Pod:"csi-node-driver-jqdjm", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief24478d807", MAC:"e2:39:d4:d1:78:8c", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:14.957950 containerd[1819]: 2025-09-13 01:22:14.956 [INFO][6300] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8" Namespace="calico-system" Pod="csi-node-driver-jqdjm" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:14.961076 containerd[1819]: time="2025-09-13T01:22:14.961055437Z" level=info msg="StartContainer for \"d3393ed809f7c32981f7b611f071196fb3620b2754431f5c81e9ce4511e8f670\" returns successfully" Sep 13 01:22:14.966654 containerd[1819]: time="2025-09-13T01:22:14.966551156Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:14.966654 containerd[1819]: time="2025-09-13T01:22:14.966583600Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:14.966654 containerd[1819]: time="2025-09-13T01:22:14.966590998Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:14.966788 containerd[1819]: time="2025-09-13T01:22:14.966640076Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:14.985536 systemd[1]: Started cri-containerd-48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8.scope - libcontainer container 48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8. Sep 13 01:22:14.995832 containerd[1819]: time="2025-09-13T01:22:14.995813097Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-jqdjm,Uid:f4d72f52-2d93-4ed1-9fa8-3baa96684daf,Namespace:calico-system,Attempt:1,} returns sandbox id \"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8\"" Sep 13 01:22:15.099229 systemd-networkd[1608]: calid074fb66ae0: Link UP Sep 13 01:22:15.099991 systemd-networkd[1608]: calid074fb66ae0: Gained carrier Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.823 [INFO][6324] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.830 [INFO][6324] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0 calico-apiserver-db75486c4- calico-apiserver 878b3bf4-9638-46e7-9a71-02b8d2591f7f 937 0 2025-09-13 01:21:47 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:db75486c4 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-n-7cd54b3887 calico-apiserver-db75486c4-sm7wq eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calid074fb66ae0 [] [] }} ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.830 [INFO][6324] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.844 [INFO][6384] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" HandleID="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.844 [INFO][6384] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" HandleID="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002e7600), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-n-7cd54b3887", "pod":"calico-apiserver-db75486c4-sm7wq", "timestamp":"2025-09-13 01:22:14.844774643 +0000 UTC"}, Hostname:"ci-4081.3.5-n-7cd54b3887", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.844 [INFO][6384] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6384] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:14.950 [INFO][6384] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-n-7cd54b3887' Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.038 [INFO][6384] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.047 [INFO][6384] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.059 [INFO][6384] ipam/ipam.go 511: Trying affinity for 192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.063 [INFO][6384] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.069 [INFO][6384] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.192/26 host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.069 [INFO][6384] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.103.192/26 handle="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.072 [INFO][6384] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.080 [INFO][6384] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.103.192/26 handle="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.091 [INFO][6384] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.103.200/26] block=192.168.103.192/26 handle="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.091 [INFO][6384] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.200/26] handle="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" host="ci-4081.3.5-n-7cd54b3887" Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.091 [INFO][6384] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:15.130969 containerd[1819]: 2025-09-13 01:22:15.091 [INFO][6384] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.103.200/26] IPv6=[] ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" HandleID="k8s-pod-network.40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:15.133237 containerd[1819]: 2025-09-13 01:22:15.095 [INFO][6324] cni-plugin/k8s.go 418: Populated endpoint ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"878b3bf4-9638-46e7-9a71-02b8d2591f7f", ResourceVersion:"937", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"", Pod:"calico-apiserver-db75486c4-sm7wq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calid074fb66ae0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:15.133237 containerd[1819]: 2025-09-13 01:22:15.096 [INFO][6324] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.200/32] ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:15.133237 containerd[1819]: 2025-09-13 01:22:15.096 [INFO][6324] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid074fb66ae0 ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:15.133237 containerd[1819]: 2025-09-13 01:22:15.099 [INFO][6324] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:15.133237 containerd[1819]: 2025-09-13 01:22:15.100 [INFO][6324] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"878b3bf4-9638-46e7-9a71-02b8d2591f7f", ResourceVersion:"937", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e", Pod:"calico-apiserver-db75486c4-sm7wq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calid074fb66ae0", MAC:"e6:f1:be:60:0f:a6", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:15.133237 containerd[1819]: 2025-09-13 01:22:15.123 [INFO][6324] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e" Namespace="calico-apiserver" Pod="calico-apiserver-db75486c4-sm7wq" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:15.144759 containerd[1819]: time="2025-09-13T01:22:15.144661430Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:22:15.144759 containerd[1819]: time="2025-09-13T01:22:15.144691308Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:22:15.144759 containerd[1819]: time="2025-09-13T01:22:15.144698434Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:15.144881 containerd[1819]: time="2025-09-13T01:22:15.144761489Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:22:15.160700 systemd[1]: Started cri-containerd-40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e.scope - libcontainer container 40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e. Sep 13 01:22:15.236239 containerd[1819]: time="2025-09-13T01:22:15.236187694Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-db75486c4-sm7wq,Uid:878b3bf4-9638-46e7-9a71-02b8d2591f7f,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e\"" Sep 13 01:22:15.542157 kubelet[3072]: I0913 01:22:15.541924 3072 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 01:22:15.814326 kernel: bpftool[6628]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Sep 13 01:22:15.866604 kubelet[3072]: I0913 01:22:15.866586 3072 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 01:22:15.889793 kubelet[3072]: I0913 01:22:15.889748 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-zkzzg" podStartSLOduration=36.889730583 podStartE2EDuration="36.889730583s" podCreationTimestamp="2025-09-13 01:21:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:22:15.889705537 +0000 UTC m=+42.214956985" watchObservedRunningTime="2025-09-13 01:22:15.889730583 +0000 UTC m=+42.214982027" Sep 13 01:22:15.981326 systemd-networkd[1608]: calif454ca88c93: Gained IPv6LL Sep 13 01:22:15.997194 systemd-networkd[1608]: vxlan.calico: Link UP Sep 13 01:22:15.997198 systemd-networkd[1608]: vxlan.calico: Gained carrier Sep 13 01:22:16.173374 systemd-networkd[1608]: cali3d80864d2e9: Gained IPv6LL Sep 13 01:22:16.621410 systemd-networkd[1608]: calid074fb66ae0: Gained IPv6LL Sep 13 01:22:16.877557 systemd-networkd[1608]: calief24478d807: Gained IPv6LL Sep 13 01:22:17.283552 containerd[1819]: time="2025-09-13T01:22:17.283462116Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:17.283766 containerd[1819]: time="2025-09-13T01:22:17.283677013Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=51277746" Sep 13 01:22:17.283974 containerd[1819]: time="2025-09-13T01:22:17.283938926Z" level=info msg="ImageCreate event name:\"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:17.285301 containerd[1819]: time="2025-09-13T01:22:17.285270108Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:17.285654 containerd[1819]: time="2025-09-13T01:22:17.285612001Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"52770417\" in 2.693141278s" Sep 13 01:22:17.285654 containerd[1819]: time="2025-09-13T01:22:17.285629463Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\"" Sep 13 01:22:17.286116 containerd[1819]: time="2025-09-13T01:22:17.286076402Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 13 01:22:17.288985 containerd[1819]: time="2025-09-13T01:22:17.288939654Z" level=info msg="CreateContainer within sandbox \"543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 13 01:22:17.293216 containerd[1819]: time="2025-09-13T01:22:17.293199519Z" level=info msg="CreateContainer within sandbox \"543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"5ec593ff253233800bae22b617dcfd89feb2b8625553cea9626cceaa5184fa2e\"" Sep 13 01:22:17.293449 containerd[1819]: time="2025-09-13T01:22:17.293435062Z" level=info msg="StartContainer for \"5ec593ff253233800bae22b617dcfd89feb2b8625553cea9626cceaa5184fa2e\"" Sep 13 01:22:17.320397 systemd[1]: Started cri-containerd-5ec593ff253233800bae22b617dcfd89feb2b8625553cea9626cceaa5184fa2e.scope - libcontainer container 5ec593ff253233800bae22b617dcfd89feb2b8625553cea9626cceaa5184fa2e. Sep 13 01:22:17.348187 containerd[1819]: time="2025-09-13T01:22:17.348158064Z" level=info msg="StartContainer for \"5ec593ff253233800bae22b617dcfd89feb2b8625553cea9626cceaa5184fa2e\" returns successfully" Sep 13 01:22:17.517507 systemd-networkd[1608]: vxlan.calico: Gained IPv6LL Sep 13 01:22:18.962572 kubelet[3072]: I0913 01:22:18.962529 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-689db9d4f-r55km" podStartSLOduration=26.633266407 podStartE2EDuration="29.962516928s" podCreationTimestamp="2025-09-13 01:21:49 +0000 UTC" firstStartedPulling="2025-09-13 01:22:13.956745416 +0000 UTC m=+40.281996861" lastFinishedPulling="2025-09-13 01:22:17.285995939 +0000 UTC m=+43.611247382" observedRunningTime="2025-09-13 01:22:17.897853722 +0000 UTC m=+44.223105239" watchObservedRunningTime="2025-09-13 01:22:18.962516928 +0000 UTC m=+45.287768377" Sep 13 01:22:19.657347 kubelet[3072]: I0913 01:22:19.657325 3072 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 01:22:19.863227 containerd[1819]: time="2025-09-13T01:22:19.863174577Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:19.863443 containerd[1819]: time="2025-09-13T01:22:19.863357140Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=47333864" Sep 13 01:22:19.863685 containerd[1819]: time="2025-09-13T01:22:19.863645190Z" level=info msg="ImageCreate event name:\"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:19.864795 containerd[1819]: time="2025-09-13T01:22:19.864756991Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:19.865206 containerd[1819]: time="2025-09-13T01:22:19.865167162Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 2.579075841s" Sep 13 01:22:19.865206 containerd[1819]: time="2025-09-13T01:22:19.865181631Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 13 01:22:19.865698 containerd[1819]: time="2025-09-13T01:22:19.865658127Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 13 01:22:19.866209 containerd[1819]: time="2025-09-13T01:22:19.866196540Z" level=info msg="CreateContainer within sandbox \"0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 13 01:22:19.869845 containerd[1819]: time="2025-09-13T01:22:19.869802828Z" level=info msg="CreateContainer within sandbox \"0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"fa457d22c5e56192a07cf7ef76c0112b1931a9e7483dd72ccef9fa7e5d38fe2c\"" Sep 13 01:22:19.870017 containerd[1819]: time="2025-09-13T01:22:19.870002478Z" level=info msg="StartContainer for \"fa457d22c5e56192a07cf7ef76c0112b1931a9e7483dd72ccef9fa7e5d38fe2c\"" Sep 13 01:22:19.889389 systemd[1]: Started cri-containerd-fa457d22c5e56192a07cf7ef76c0112b1931a9e7483dd72ccef9fa7e5d38fe2c.scope - libcontainer container fa457d22c5e56192a07cf7ef76c0112b1931a9e7483dd72ccef9fa7e5d38fe2c. Sep 13 01:22:19.912777 containerd[1819]: time="2025-09-13T01:22:19.912712437Z" level=info msg="StartContainer for \"fa457d22c5e56192a07cf7ef76c0112b1931a9e7483dd72ccef9fa7e5d38fe2c\" returns successfully" Sep 13 01:22:20.892328 kubelet[3072]: I0913 01:22:20.892291 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-db75486c4-rqw44" podStartSLOduration=28.064607616 podStartE2EDuration="33.89227883s" podCreationTimestamp="2025-09-13 01:21:47 +0000 UTC" firstStartedPulling="2025-09-13 01:22:14.037933553 +0000 UTC m=+40.363184996" lastFinishedPulling="2025-09-13 01:22:19.865604765 +0000 UTC m=+46.190856210" observedRunningTime="2025-09-13 01:22:20.892070623 +0000 UTC m=+47.217322067" watchObservedRunningTime="2025-09-13 01:22:20.89227883 +0000 UTC m=+47.217530271" Sep 13 01:22:21.480517 containerd[1819]: time="2025-09-13T01:22:21.480478355Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:21.480843 containerd[1819]: time="2025-09-13T01:22:21.480683347Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8760527" Sep 13 01:22:21.481090 containerd[1819]: time="2025-09-13T01:22:21.481074115Z" level=info msg="ImageCreate event name:\"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:21.482112 containerd[1819]: time="2025-09-13T01:22:21.482077077Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:21.482566 containerd[1819]: time="2025-09-13T01:22:21.482524209Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"10253230\" in 1.616851641s" Sep 13 01:22:21.482566 containerd[1819]: time="2025-09-13T01:22:21.482565606Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\"" Sep 13 01:22:21.483213 containerd[1819]: time="2025-09-13T01:22:21.483202085Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 13 01:22:21.483830 containerd[1819]: time="2025-09-13T01:22:21.483818150Z" level=info msg="CreateContainer within sandbox \"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 13 01:22:21.489002 containerd[1819]: time="2025-09-13T01:22:21.488929723Z" level=info msg="CreateContainer within sandbox \"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"84870e71a19eb9135d6d7de0e33ee88461c01c49c031520e8eb6f8c86325f4be\"" Sep 13 01:22:21.489201 containerd[1819]: time="2025-09-13T01:22:21.489189944Z" level=info msg="StartContainer for \"84870e71a19eb9135d6d7de0e33ee88461c01c49c031520e8eb6f8c86325f4be\"" Sep 13 01:22:21.523552 systemd[1]: Started cri-containerd-84870e71a19eb9135d6d7de0e33ee88461c01c49c031520e8eb6f8c86325f4be.scope - libcontainer container 84870e71a19eb9135d6d7de0e33ee88461c01c49c031520e8eb6f8c86325f4be. Sep 13 01:22:21.538260 containerd[1819]: time="2025-09-13T01:22:21.538220355Z" level=info msg="StartContainer for \"84870e71a19eb9135d6d7de0e33ee88461c01c49c031520e8eb6f8c86325f4be\" returns successfully" Sep 13 01:22:21.921425 containerd[1819]: time="2025-09-13T01:22:21.921404288Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:21.921718 containerd[1819]: time="2025-09-13T01:22:21.921667821Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=77" Sep 13 01:22:21.922920 containerd[1819]: time="2025-09-13T01:22:21.922879136Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 439.662212ms" Sep 13 01:22:21.922920 containerd[1819]: time="2025-09-13T01:22:21.922894451Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 13 01:22:21.923416 containerd[1819]: time="2025-09-13T01:22:21.923381868Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 13 01:22:21.923963 containerd[1819]: time="2025-09-13T01:22:21.923949470Z" level=info msg="CreateContainer within sandbox \"40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 13 01:22:21.928528 containerd[1819]: time="2025-09-13T01:22:21.928484388Z" level=info msg="CreateContainer within sandbox \"40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"2782d1031159f66b710225860fb9bd4042ac775b1033d099f543652cd7c67a48\"" Sep 13 01:22:21.928851 containerd[1819]: time="2025-09-13T01:22:21.928775409Z" level=info msg="StartContainer for \"2782d1031159f66b710225860fb9bd4042ac775b1033d099f543652cd7c67a48\"" Sep 13 01:22:21.951548 systemd[1]: Started cri-containerd-2782d1031159f66b710225860fb9bd4042ac775b1033d099f543652cd7c67a48.scope - libcontainer container 2782d1031159f66b710225860fb9bd4042ac775b1033d099f543652cd7c67a48. Sep 13 01:22:21.980345 containerd[1819]: time="2025-09-13T01:22:21.980315270Z" level=info msg="StartContainer for \"2782d1031159f66b710225860fb9bd4042ac775b1033d099f543652cd7c67a48\" returns successfully" Sep 13 01:22:22.902488 kubelet[3072]: I0913 01:22:22.902456 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-db75486c4-sm7wq" podStartSLOduration=29.216654692 podStartE2EDuration="35.902445101s" podCreationTimestamp="2025-09-13 01:21:47 +0000 UTC" firstStartedPulling="2025-09-13 01:22:15.23746444 +0000 UTC m=+41.562715908" lastFinishedPulling="2025-09-13 01:22:21.923254874 +0000 UTC m=+48.248506317" observedRunningTime="2025-09-13 01:22:22.902430293 +0000 UTC m=+49.227681738" watchObservedRunningTime="2025-09-13 01:22:22.902445101 +0000 UTC m=+49.227696542" Sep 13 01:22:23.741924 containerd[1819]: time="2025-09-13T01:22:23.741901223Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:23.742143 containerd[1819]: time="2025-09-13T01:22:23.742095210Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=14698542" Sep 13 01:22:23.742443 containerd[1819]: time="2025-09-13T01:22:23.742429801Z" level=info msg="ImageCreate event name:\"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:23.744003 containerd[1819]: time="2025-09-13T01:22:23.743990565Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:22:23.744314 containerd[1819]: time="2025-09-13T01:22:23.744254469Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"16191197\" in 1.82084977s" Sep 13 01:22:23.744314 containerd[1819]: time="2025-09-13T01:22:23.744268820Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\"" Sep 13 01:22:23.745195 containerd[1819]: time="2025-09-13T01:22:23.745180563Z" level=info msg="CreateContainer within sandbox \"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 13 01:22:23.749546 containerd[1819]: time="2025-09-13T01:22:23.749529578Z" level=info msg="CreateContainer within sandbox \"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"13e0a733234454003cc913bc89b7a49418280e5bf0f5dacdc66899f6320ebf6e\"" Sep 13 01:22:23.749743 containerd[1819]: time="2025-09-13T01:22:23.749729879Z" level=info msg="StartContainer for \"13e0a733234454003cc913bc89b7a49418280e5bf0f5dacdc66899f6320ebf6e\"" Sep 13 01:22:23.780689 systemd[1]: Started cri-containerd-13e0a733234454003cc913bc89b7a49418280e5bf0f5dacdc66899f6320ebf6e.scope - libcontainer container 13e0a733234454003cc913bc89b7a49418280e5bf0f5dacdc66899f6320ebf6e. Sep 13 01:22:23.801347 containerd[1819]: time="2025-09-13T01:22:23.801292545Z" level=info msg="StartContainer for \"13e0a733234454003cc913bc89b7a49418280e5bf0f5dacdc66899f6320ebf6e\" returns successfully" Sep 13 01:22:23.900685 kubelet[3072]: I0913 01:22:23.900651 3072 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 01:22:24.764705 kubelet[3072]: I0913 01:22:24.764608 3072 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 13 01:22:24.764705 kubelet[3072]: I0913 01:22:24.764676 3072 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 13 01:22:33.718792 containerd[1819]: time="2025-09-13T01:22:33.718596583Z" level=info msg="StopPodSandbox for \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\"" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.751 [WARNING][7212] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0", GenerateName:"calico-kube-controllers-689db9d4f-", Namespace:"calico-system", SelfLink:"", UID:"198f178f-c689-48cb-b5e4-012f47ac61f3", ResourceVersion:"984", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"689db9d4f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03", Pod:"calico-kube-controllers-689db9d4f-r55km", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calia435cbd0940", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.751 [INFO][7212] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.751 [INFO][7212] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" iface="eth0" netns="" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.751 [INFO][7212] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.751 [INFO][7212] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.761 [INFO][7230] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.761 [INFO][7230] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.761 [INFO][7230] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.765 [WARNING][7230] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.765 [INFO][7230] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.766 [INFO][7230] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:33.768394 containerd[1819]: 2025-09-13 01:22:33.767 [INFO][7212] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.768394 containerd[1819]: time="2025-09-13T01:22:33.768383485Z" level=info msg="TearDown network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\" successfully" Sep 13 01:22:33.768394 containerd[1819]: time="2025-09-13T01:22:33.768402078Z" level=info msg="StopPodSandbox for \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\" returns successfully" Sep 13 01:22:33.768729 containerd[1819]: time="2025-09-13T01:22:33.768691603Z" level=info msg="RemovePodSandbox for \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\"" Sep 13 01:22:33.768729 containerd[1819]: time="2025-09-13T01:22:33.768710527Z" level=info msg="Forcibly stopping sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\"" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.792 [WARNING][7257] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0", GenerateName:"calico-kube-controllers-689db9d4f-", Namespace:"calico-system", SelfLink:"", UID:"198f178f-c689-48cb-b5e4-012f47ac61f3", ResourceVersion:"984", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"689db9d4f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"543eabd52ec7867eafa33d5beae411bb460c49f00180ed9a43c30c95c2ac5b03", Pod:"calico-kube-controllers-689db9d4f-r55km", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calia435cbd0940", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.793 [INFO][7257] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.793 [INFO][7257] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" iface="eth0" netns="" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.793 [INFO][7257] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.793 [INFO][7257] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.808 [INFO][7271] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.808 [INFO][7271] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.808 [INFO][7271] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.812 [WARNING][7271] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.812 [INFO][7271] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" HandleID="k8s-pod-network.5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--kube--controllers--689db9d4f--r55km-eth0" Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.813 [INFO][7271] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:33.815085 containerd[1819]: 2025-09-13 01:22:33.814 [INFO][7257] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272" Sep 13 01:22:33.815422 containerd[1819]: time="2025-09-13T01:22:33.815111519Z" level=info msg="TearDown network for sandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\" successfully" Sep 13 01:22:33.816599 containerd[1819]: time="2025-09-13T01:22:33.816559012Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:33.816599 containerd[1819]: time="2025-09-13T01:22:33.816593820Z" level=info msg="RemovePodSandbox \"5b35f69d186a0ab461394d0a2cc96bc24c9bcefba3aafccfeb31ff6bbb876272\" returns successfully" Sep 13 01:22:33.816942 containerd[1819]: time="2025-09-13T01:22:33.816912208Z" level=info msg="StopPodSandbox for \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\"" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.834 [WARNING][7295] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"3999b298-1754-45b1-95dc-2f560b55d1e4", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d", Pod:"goldmane-54d579b49d-sfpqb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calibfbc197f5b1", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.834 [INFO][7295] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.834 [INFO][7295] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" iface="eth0" netns="" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.834 [INFO][7295] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.834 [INFO][7295] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.843 [INFO][7314] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.843 [INFO][7314] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.843 [INFO][7314] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.847 [WARNING][7314] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.847 [INFO][7314] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.848 [INFO][7314] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:33.849548 containerd[1819]: 2025-09-13 01:22:33.848 [INFO][7295] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.849548 containerd[1819]: time="2025-09-13T01:22:33.849514307Z" level=info msg="TearDown network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\" successfully" Sep 13 01:22:33.849548 containerd[1819]: time="2025-09-13T01:22:33.849530440Z" level=info msg="StopPodSandbox for \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\" returns successfully" Sep 13 01:22:33.849876 containerd[1819]: time="2025-09-13T01:22:33.849806583Z" level=info msg="RemovePodSandbox for \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\"" Sep 13 01:22:33.849876 containerd[1819]: time="2025-09-13T01:22:33.849824225Z" level=info msg="Forcibly stopping sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\"" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.867 [WARNING][7339] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"3999b298-1754-45b1-95dc-2f560b55d1e4", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"a902e7c475472db7c16653ac9c4b1d20a21555aaf82d64ca4f0de7b19cd2453d", Pod:"goldmane-54d579b49d-sfpqb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calibfbc197f5b1", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.868 [INFO][7339] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.868 [INFO][7339] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" iface="eth0" netns="" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.868 [INFO][7339] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.868 [INFO][7339] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.878 [INFO][7355] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.878 [INFO][7355] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.878 [INFO][7355] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.881 [WARNING][7355] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.882 [INFO][7355] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" HandleID="k8s-pod-network.de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Workload="ci--4081.3.5--n--7cd54b3887-k8s-goldmane--54d579b49d--sfpqb-eth0" Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.882 [INFO][7355] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:33.884435 containerd[1819]: 2025-09-13 01:22:33.883 [INFO][7339] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7" Sep 13 01:22:33.884742 containerd[1819]: time="2025-09-13T01:22:33.884463780Z" level=info msg="TearDown network for sandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\" successfully" Sep 13 01:22:33.903042 containerd[1819]: time="2025-09-13T01:22:33.902999091Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:33.903042 containerd[1819]: time="2025-09-13T01:22:33.903032987Z" level=info msg="RemovePodSandbox \"de63bf951ce456fd8222a73d5ed10027f44ba3daffcb5514deecc32d72ee10b7\" returns successfully" Sep 13 01:22:33.903287 containerd[1819]: time="2025-09-13T01:22:33.903275455Z" level=info msg="StopPodSandbox for \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\"" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.920 [WARNING][7381] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f4d72f52-2d93-4ed1-9fa8-3baa96684daf", ResourceVersion:"1032", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8", Pod:"csi-node-driver-jqdjm", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief24478d807", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.921 [INFO][7381] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.921 [INFO][7381] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" iface="eth0" netns="" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.921 [INFO][7381] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.921 [INFO][7381] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.930 [INFO][7399] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.931 [INFO][7399] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.931 [INFO][7399] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.934 [WARNING][7399] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.934 [INFO][7399] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.935 [INFO][7399] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:33.937302 containerd[1819]: 2025-09-13 01:22:33.936 [INFO][7381] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.937599 containerd[1819]: time="2025-09-13T01:22:33.937300564Z" level=info msg="TearDown network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\" successfully" Sep 13 01:22:33.937599 containerd[1819]: time="2025-09-13T01:22:33.937316225Z" level=info msg="StopPodSandbox for \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\" returns successfully" Sep 13 01:22:33.937599 containerd[1819]: time="2025-09-13T01:22:33.937564703Z" level=info msg="RemovePodSandbox for \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\"" Sep 13 01:22:33.937599 containerd[1819]: time="2025-09-13T01:22:33.937579997Z" level=info msg="Forcibly stopping sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\"" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.955 [WARNING][7421] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f4d72f52-2d93-4ed1-9fa8-3baa96684daf", ResourceVersion:"1032", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"48fd50c56030fd18b5e7760dc58822c55a49f4799e650ade87ba500cd1a692b8", Pod:"csi-node-driver-jqdjm", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calief24478d807", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.955 [INFO][7421] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.955 [INFO][7421] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" iface="eth0" netns="" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.955 [INFO][7421] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.955 [INFO][7421] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.965 [INFO][7438] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.965 [INFO][7438] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.965 [INFO][7438] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.970 [WARNING][7438] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.970 [INFO][7438] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" HandleID="k8s-pod-network.0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Workload="ci--4081.3.5--n--7cd54b3887-k8s-csi--node--driver--jqdjm-eth0" Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.971 [INFO][7438] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:33.973058 containerd[1819]: 2025-09-13 01:22:33.972 [INFO][7421] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6" Sep 13 01:22:33.973058 containerd[1819]: time="2025-09-13T01:22:33.973014365Z" level=info msg="TearDown network for sandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\" successfully" Sep 13 01:22:33.974644 containerd[1819]: time="2025-09-13T01:22:33.974601003Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:33.974644 containerd[1819]: time="2025-09-13T01:22:33.974638197Z" level=info msg="RemovePodSandbox \"0808a5e33fa8aadd4de98ce0c96923b05c4bd7b4485fc77e4c5c570912d0d0b6\" returns successfully" Sep 13 01:22:33.974904 containerd[1819]: time="2025-09-13T01:22:33.974893342Z" level=info msg="StopPodSandbox for \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\"" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:33.992 [WARNING][7464] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:33.992 [INFO][7464] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:33.992 [INFO][7464] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" iface="eth0" netns="" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:33.992 [INFO][7464] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:33.992 [INFO][7464] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:34.002 [INFO][7478] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:34.002 [INFO][7478] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:34.002 [INFO][7478] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:34.006 [WARNING][7478] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:34.006 [INFO][7478] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:34.007 [INFO][7478] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.009297 containerd[1819]: 2025-09-13 01:22:34.008 [INFO][7464] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.009601 containerd[1819]: time="2025-09-13T01:22:34.009320401Z" level=info msg="TearDown network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\" successfully" Sep 13 01:22:34.009601 containerd[1819]: time="2025-09-13T01:22:34.009337475Z" level=info msg="StopPodSandbox for \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\" returns successfully" Sep 13 01:22:34.009651 containerd[1819]: time="2025-09-13T01:22:34.009606009Z" level=info msg="RemovePodSandbox for \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\"" Sep 13 01:22:34.009651 containerd[1819]: time="2025-09-13T01:22:34.009627592Z" level=info msg="Forcibly stopping sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\"" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.026 [WARNING][7505] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" WorkloadEndpoint="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.026 [INFO][7505] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.026 [INFO][7505] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" iface="eth0" netns="" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.026 [INFO][7505] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.026 [INFO][7505] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.036 [INFO][7522] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.036 [INFO][7522] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.036 [INFO][7522] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.040 [WARNING][7522] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.040 [INFO][7522] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" HandleID="k8s-pod-network.9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Workload="ci--4081.3.5--n--7cd54b3887-k8s-whisker--85749946d--xvsn5-eth0" Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.041 [INFO][7522] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.043388 containerd[1819]: 2025-09-13 01:22:34.042 [INFO][7505] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861" Sep 13 01:22:34.043388 containerd[1819]: time="2025-09-13T01:22:34.043382823Z" level=info msg="TearDown network for sandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\" successfully" Sep 13 01:22:34.044804 containerd[1819]: time="2025-09-13T01:22:34.044760152Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:34.044804 containerd[1819]: time="2025-09-13T01:22:34.044792593Z" level=info msg="RemovePodSandbox \"9977441d91b7639afd54afdcbb199addae58c2523934ac204237d47631620861\" returns successfully" Sep 13 01:22:34.045100 containerd[1819]: time="2025-09-13T01:22:34.045062102Z" level=info msg="StopPodSandbox for \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\"" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.067 [WARNING][7547] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1e417130-1b3b-4716-84aa-d920f0d71e2e", ResourceVersion:"1004", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1", Pod:"calico-apiserver-db75486c4-rqw44", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif454ca88c93", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.067 [INFO][7547] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.067 [INFO][7547] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" iface="eth0" netns="" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.067 [INFO][7547] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.068 [INFO][7547] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.105 [INFO][7565] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.105 [INFO][7565] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.105 [INFO][7565] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.111 [WARNING][7565] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.111 [INFO][7565] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.113 [INFO][7565] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.115572 containerd[1819]: 2025-09-13 01:22:34.114 [INFO][7547] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.116055 containerd[1819]: time="2025-09-13T01:22:34.115607035Z" level=info msg="TearDown network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\" successfully" Sep 13 01:22:34.116055 containerd[1819]: time="2025-09-13T01:22:34.115633601Z" level=info msg="StopPodSandbox for \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\" returns successfully" Sep 13 01:22:34.116055 containerd[1819]: time="2025-09-13T01:22:34.116031425Z" level=info msg="RemovePodSandbox for \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\"" Sep 13 01:22:34.116376 containerd[1819]: time="2025-09-13T01:22:34.116062534Z" level=info msg="Forcibly stopping sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\"" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.144 [WARNING][7591] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1e417130-1b3b-4716-84aa-d920f0d71e2e", ResourceVersion:"1004", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"0873d2d15f3592e12d1e6e64ab69d584ee8062e8334678ea138f4ad925232be1", Pod:"calico-apiserver-db75486c4-rqw44", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif454ca88c93", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.145 [INFO][7591] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.145 [INFO][7591] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" iface="eth0" netns="" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.145 [INFO][7591] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.145 [INFO][7591] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.162 [INFO][7607] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.162 [INFO][7607] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.162 [INFO][7607] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.168 [WARNING][7607] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.169 [INFO][7607] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" HandleID="k8s-pod-network.5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--rqw44-eth0" Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.170 [INFO][7607] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.172853 containerd[1819]: 2025-09-13 01:22:34.171 [INFO][7591] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911" Sep 13 01:22:34.173358 containerd[1819]: time="2025-09-13T01:22:34.172869977Z" level=info msg="TearDown network for sandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\" successfully" Sep 13 01:22:34.174836 containerd[1819]: time="2025-09-13T01:22:34.174794685Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:34.174867 containerd[1819]: time="2025-09-13T01:22:34.174843946Z" level=info msg="RemovePodSandbox \"5dff8e775003fd328964b9cb2c4625542812c10d0ee8774a93799f7b04a32911\" returns successfully" Sep 13 01:22:34.175139 containerd[1819]: time="2025-09-13T01:22:34.175094394Z" level=info msg="StopPodSandbox for \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\"" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.193 [WARNING][7630] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0db74307-721a-4439-bd8d-982b96aa7db8", ResourceVersion:"970", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc", Pod:"coredns-668d6bf9bc-zkzzg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3d80864d2e9", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.193 [INFO][7630] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.193 [INFO][7630] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" iface="eth0" netns="" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.193 [INFO][7630] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.193 [INFO][7630] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.204 [INFO][7649] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.204 [INFO][7649] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.204 [INFO][7649] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.208 [WARNING][7649] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.208 [INFO][7649] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.210 [INFO][7649] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.211643 containerd[1819]: 2025-09-13 01:22:34.210 [INFO][7630] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.211976 containerd[1819]: time="2025-09-13T01:22:34.211652918Z" level=info msg="TearDown network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\" successfully" Sep 13 01:22:34.211976 containerd[1819]: time="2025-09-13T01:22:34.211669531Z" level=info msg="StopPodSandbox for \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\" returns successfully" Sep 13 01:22:34.211976 containerd[1819]: time="2025-09-13T01:22:34.211924552Z" level=info msg="RemovePodSandbox for \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\"" Sep 13 01:22:34.211976 containerd[1819]: time="2025-09-13T01:22:34.211943277Z" level=info msg="Forcibly stopping sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\"" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.230 [WARNING][7675] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0db74307-721a-4439-bd8d-982b96aa7db8", ResourceVersion:"970", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"8596d2f9aac7f0f93e6162582fcccf148b486cb67d3b602c68e05f2a947bc5bc", Pod:"coredns-668d6bf9bc-zkzzg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3d80864d2e9", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.231 [INFO][7675] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.231 [INFO][7675] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" iface="eth0" netns="" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.231 [INFO][7675] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.231 [INFO][7675] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.243 [INFO][7690] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.243 [INFO][7690] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.243 [INFO][7690] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.248 [WARNING][7690] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.248 [INFO][7690] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" HandleID="k8s-pod-network.c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--zkzzg-eth0" Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.249 [INFO][7690] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.251412 containerd[1819]: 2025-09-13 01:22:34.250 [INFO][7675] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07" Sep 13 01:22:34.251412 containerd[1819]: time="2025-09-13T01:22:34.251395987Z" level=info msg="TearDown network for sandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\" successfully" Sep 13 01:22:34.253271 containerd[1819]: time="2025-09-13T01:22:34.253225476Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:34.253271 containerd[1819]: time="2025-09-13T01:22:34.253261815Z" level=info msg="RemovePodSandbox \"c2a7f755cda832a5dc13b9915b637015efa237e1350191faef20eb95a6e74a07\" returns successfully" Sep 13 01:22:34.253553 containerd[1819]: time="2025-09-13T01:22:34.253511385Z" level=info msg="StopPodSandbox for \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\"" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.271 [WARNING][7714] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"878b3bf4-9638-46e7-9a71-02b8d2591f7f", ResourceVersion:"1023", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e", Pod:"calico-apiserver-db75486c4-sm7wq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calid074fb66ae0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.271 [INFO][7714] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.271 [INFO][7714] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" iface="eth0" netns="" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.271 [INFO][7714] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.271 [INFO][7714] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.284 [INFO][7729] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.284 [INFO][7729] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.284 [INFO][7729] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.288 [WARNING][7729] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.288 [INFO][7729] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.289 [INFO][7729] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.291431 containerd[1819]: 2025-09-13 01:22:34.290 [INFO][7714] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.291431 containerd[1819]: time="2025-09-13T01:22:34.291425683Z" level=info msg="TearDown network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\" successfully" Sep 13 01:22:34.291787 containerd[1819]: time="2025-09-13T01:22:34.291445344Z" level=info msg="StopPodSandbox for \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\" returns successfully" Sep 13 01:22:34.291810 containerd[1819]: time="2025-09-13T01:22:34.291786182Z" level=info msg="RemovePodSandbox for \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\"" Sep 13 01:22:34.291810 containerd[1819]: time="2025-09-13T01:22:34.291805471Z" level=info msg="Forcibly stopping sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\"" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.312 [WARNING][7751] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0", GenerateName:"calico-apiserver-db75486c4-", Namespace:"calico-apiserver", SelfLink:"", UID:"878b3bf4-9638-46e7-9a71-02b8d2591f7f", ResourceVersion:"1023", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"db75486c4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"40c9cf74914b9194770ac9fc557f883543f7a6d9f6d836fae60ed50c26f3156e", Pod:"calico-apiserver-db75486c4-sm7wq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calid074fb66ae0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.312 [INFO][7751] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.312 [INFO][7751] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" iface="eth0" netns="" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.312 [INFO][7751] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.312 [INFO][7751] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.323 [INFO][7765] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.323 [INFO][7765] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.323 [INFO][7765] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.326 [WARNING][7765] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.326 [INFO][7765] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" HandleID="k8s-pod-network.08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Workload="ci--4081.3.5--n--7cd54b3887-k8s-calico--apiserver--db75486c4--sm7wq-eth0" Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.327 [INFO][7765] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.329374 containerd[1819]: 2025-09-13 01:22:34.328 [INFO][7751] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b" Sep 13 01:22:34.329374 containerd[1819]: time="2025-09-13T01:22:34.329373524Z" level=info msg="TearDown network for sandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\" successfully" Sep 13 01:22:34.330755 containerd[1819]: time="2025-09-13T01:22:34.330714274Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:34.330755 containerd[1819]: time="2025-09-13T01:22:34.330743813Z" level=info msg="RemovePodSandbox \"08c635ba8e78264c9fc9847a25ea6197bbc6bb9e1a1cf259942de58c40f2295b\" returns successfully" Sep 13 01:22:34.331013 containerd[1819]: time="2025-09-13T01:22:34.330983747Z" level=info msg="StopPodSandbox for \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\"" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.348 [WARNING][7791] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"465526ae-5203-429b-8157-1675929b92b0", ResourceVersion:"908", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b", Pod:"coredns-668d6bf9bc-5hgc5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia3c243e27ae", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.348 [INFO][7791] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.348 [INFO][7791] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" iface="eth0" netns="" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.348 [INFO][7791] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.348 [INFO][7791] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.359 [INFO][7808] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.359 [INFO][7808] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.359 [INFO][7808] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.363 [WARNING][7808] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.363 [INFO][7808] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.364 [INFO][7808] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.366109 containerd[1819]: 2025-09-13 01:22:34.365 [INFO][7791] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.366501 containerd[1819]: time="2025-09-13T01:22:34.366136814Z" level=info msg="TearDown network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\" successfully" Sep 13 01:22:34.366501 containerd[1819]: time="2025-09-13T01:22:34.366153513Z" level=info msg="StopPodSandbox for \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\" returns successfully" Sep 13 01:22:34.366501 containerd[1819]: time="2025-09-13T01:22:34.366432822Z" level=info msg="RemovePodSandbox for \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\"" Sep 13 01:22:34.366501 containerd[1819]: time="2025-09-13T01:22:34.366449278Z" level=info msg="Forcibly stopping sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\"" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.385 [WARNING][7834] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"465526ae-5203-429b-8157-1675929b92b0", ResourceVersion:"908", Generation:0, CreationTimestamp:time.Date(2025, time.September, 13, 1, 21, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-n-7cd54b3887", ContainerID:"969da10a01391317ebe9c7f1cd47d9b59b33e38a7434f56fd0ec092c0d965f7b", Pod:"coredns-668d6bf9bc-5hgc5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia3c243e27ae", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.386 [INFO][7834] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.386 [INFO][7834] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" iface="eth0" netns="" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.386 [INFO][7834] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.386 [INFO][7834] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.396 [INFO][7849] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.397 [INFO][7849] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.397 [INFO][7849] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.401 [WARNING][7849] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.401 [INFO][7849] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" HandleID="k8s-pod-network.6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Workload="ci--4081.3.5--n--7cd54b3887-k8s-coredns--668d6bf9bc--5hgc5-eth0" Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.402 [INFO][7849] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 13 01:22:34.404090 containerd[1819]: 2025-09-13 01:22:34.403 [INFO][7834] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba" Sep 13 01:22:34.404442 containerd[1819]: time="2025-09-13T01:22:34.404116569Z" level=info msg="TearDown network for sandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\" successfully" Sep 13 01:22:34.405654 containerd[1819]: time="2025-09-13T01:22:34.405612425Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 13 01:22:34.405654 containerd[1819]: time="2025-09-13T01:22:34.405647505Z" level=info msg="RemovePodSandbox \"6037697239ac7a7a20b1a2a32f629a36e48dd094820252ac923000032f72eeba\" returns successfully" Sep 13 01:22:36.904223 kubelet[3072]: I0913 01:22:36.904193 3072 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-jqdjm" podStartSLOduration=39.155900286 podStartE2EDuration="47.904183993s" podCreationTimestamp="2025-09-13 01:21:49 +0000 UTC" firstStartedPulling="2025-09-13 01:22:14.996343053 +0000 UTC m=+41.321594496" lastFinishedPulling="2025-09-13 01:22:23.744626763 +0000 UTC m=+50.069878203" observedRunningTime="2025-09-13 01:22:23.91878491 +0000 UTC m=+50.244036436" watchObservedRunningTime="2025-09-13 01:22:36.904183993 +0000 UTC m=+63.229435433" Sep 13 01:22:37.018295 kubelet[3072]: I0913 01:22:37.018185 3072 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 01:26:48.116742 systemd[1]: Started sshd@9-139.178.94.15:22-185.156.73.233:16588.service - OpenSSH per-connection server daemon (185.156.73.233:16588). Sep 13 01:26:50.326829 sshd[8871]: Invalid user ubnt from 185.156.73.233 port 16588 Sep 13 01:26:50.473895 sshd[8871]: Connection closed by invalid user ubnt 185.156.73.233 port 16588 [preauth] Sep 13 01:26:50.478862 systemd[1]: sshd@9-139.178.94.15:22-185.156.73.233:16588.service: Deactivated successfully. Sep 13 01:27:34.197977 systemd[1]: Started sshd@10-139.178.94.15:22-139.178.89.65:53080.service - OpenSSH per-connection server daemon (139.178.89.65:53080). Sep 13 01:27:34.264790 sshd[9092]: Accepted publickey for core from 139.178.89.65 port 53080 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:27:34.265861 sshd[9092]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:27:34.269541 systemd-logind[1809]: New session 12 of user core. Sep 13 01:27:34.281437 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 13 01:27:34.414827 sshd[9092]: pam_unix(sshd:session): session closed for user core Sep 13 01:27:34.416228 systemd[1]: sshd@10-139.178.94.15:22-139.178.89.65:53080.service: Deactivated successfully. Sep 13 01:27:34.417125 systemd[1]: session-12.scope: Deactivated successfully. Sep 13 01:27:34.417749 systemd-logind[1809]: Session 12 logged out. Waiting for processes to exit. Sep 13 01:27:34.418195 systemd-logind[1809]: Removed session 12. Sep 13 01:27:39.457493 systemd[1]: Started sshd@11-139.178.94.15:22-139.178.89.65:53090.service - OpenSSH per-connection server daemon (139.178.89.65:53090). Sep 13 01:27:39.492536 sshd[9166]: Accepted publickey for core from 139.178.89.65 port 53090 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:27:39.493418 sshd[9166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:27:39.496200 systemd-logind[1809]: New session 13 of user core. Sep 13 01:27:39.514414 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 13 01:27:39.635520 sshd[9166]: pam_unix(sshd:session): session closed for user core Sep 13 01:27:39.637084 systemd[1]: sshd@11-139.178.94.15:22-139.178.89.65:53090.service: Deactivated successfully. Sep 13 01:27:39.638036 systemd[1]: session-13.scope: Deactivated successfully. Sep 13 01:27:39.638763 systemd-logind[1809]: Session 13 logged out. Waiting for processes to exit. Sep 13 01:27:39.639240 systemd-logind[1809]: Removed session 13. Sep 13 01:27:44.677479 systemd[1]: Started sshd@12-139.178.94.15:22-139.178.89.65:33090.service - OpenSSH per-connection server daemon (139.178.89.65:33090). Sep 13 01:27:44.707163 sshd[9195]: Accepted publickey for core from 139.178.89.65 port 33090 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:27:44.707900 sshd[9195]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:27:44.710686 systemd-logind[1809]: New session 14 of user core. Sep 13 01:27:44.718531 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 13 01:27:44.805224 sshd[9195]: pam_unix(sshd:session): session closed for user core Sep 13 01:27:44.824400 systemd[1]: sshd@12-139.178.94.15:22-139.178.89.65:33090.service: Deactivated successfully. Sep 13 01:27:44.825529 systemd[1]: session-14.scope: Deactivated successfully. Sep 13 01:27:44.826388 systemd-logind[1809]: Session 14 logged out. Waiting for processes to exit. Sep 13 01:27:44.827362 systemd[1]: Started sshd@13-139.178.94.15:22-139.178.89.65:33104.service - OpenSSH per-connection server daemon (139.178.89.65:33104). Sep 13 01:27:44.828002 systemd-logind[1809]: Removed session 14. Sep 13 01:27:44.884428 sshd[9222]: Accepted publickey for core from 139.178.89.65 port 33104 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:27:44.885659 sshd[9222]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:27:44.889315 systemd-logind[1809]: New session 15 of user core. Sep 13 01:27:44.914511 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 13 01:27:45.052585 sshd[9222]: pam_unix(sshd:session): session closed for user core Sep 13 01:27:45.063921 systemd[1]: sshd@13-139.178.94.15:22-139.178.89.65:33104.service: Deactivated successfully. Sep 13 01:27:45.064873 systemd[1]: session-15.scope: Deactivated successfully. Sep 13 01:27:45.065561 systemd-logind[1809]: Session 15 logged out. Waiting for processes to exit. Sep 13 01:27:45.066234 systemd[1]: Started sshd@14-139.178.94.15:22-139.178.89.65:33118.service - OpenSSH per-connection server daemon (139.178.89.65:33118). Sep 13 01:27:45.066661 systemd-logind[1809]: Removed session 15. Sep 13 01:27:45.096691 sshd[9246]: Accepted publickey for core from 139.178.89.65 port 33118 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:27:45.097377 sshd[9246]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:27:45.099897 systemd-logind[1809]: New session 16 of user core. Sep 13 01:27:45.118446 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 13 01:27:45.263219 sshd[9246]: pam_unix(sshd:session): session closed for user core Sep 13 01:27:45.264975 systemd[1]: sshd@14-139.178.94.15:22-139.178.89.65:33118.service: Deactivated successfully. Sep 13 01:27:45.266035 systemd[1]: session-16.scope: Deactivated successfully. Sep 13 01:27:45.266910 systemd-logind[1809]: Session 16 logged out. Waiting for processes to exit. Sep 13 01:27:45.267738 systemd-logind[1809]: Removed session 16. Sep 13 01:27:50.290177 systemd[1]: Started sshd@15-139.178.94.15:22-139.178.89.65:38834.service - OpenSSH per-connection server daemon (139.178.89.65:38834). Sep 13 01:27:50.326743 sshd[9325]: Accepted publickey for core from 139.178.89.65 port 38834 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:27:50.327463 sshd[9325]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:27:50.329957 systemd-logind[1809]: New session 17 of user core. Sep 13 01:27:50.341561 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 13 01:27:50.430502 sshd[9325]: pam_unix(sshd:session): session closed for user core Sep 13 01:27:50.432138 systemd[1]: sshd@15-139.178.94.15:22-139.178.89.65:38834.service: Deactivated successfully. Sep 13 01:27:50.433120 systemd[1]: session-17.scope: Deactivated successfully. Sep 13 01:27:50.433878 systemd-logind[1809]: Session 17 logged out. Waiting for processes to exit. Sep 13 01:27:50.434468 systemd-logind[1809]: Removed session 17. Sep 13 01:27:55.451159 systemd[1]: Started sshd@16-139.178.94.15:22-139.178.89.65:38840.service - OpenSSH per-connection server daemon (139.178.89.65:38840). Sep 13 01:27:55.481604 sshd[9351]: Accepted publickey for core from 139.178.89.65 port 38840 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:27:55.482299 sshd[9351]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:27:55.484865 systemd-logind[1809]: New session 18 of user core. Sep 13 01:27:55.485451 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 13 01:27:55.570106 sshd[9351]: pam_unix(sshd:session): session closed for user core Sep 13 01:27:55.571808 systemd[1]: sshd@16-139.178.94.15:22-139.178.89.65:38840.service: Deactivated successfully. Sep 13 01:27:55.572780 systemd[1]: session-18.scope: Deactivated successfully. Sep 13 01:27:55.573584 systemd-logind[1809]: Session 18 logged out. Waiting for processes to exit. Sep 13 01:27:55.574175 systemd-logind[1809]: Removed session 18. Sep 13 01:28:00.593193 systemd[1]: Started sshd@17-139.178.94.15:22-139.178.89.65:35632.service - OpenSSH per-connection server daemon (139.178.89.65:35632). Sep 13 01:28:00.623817 sshd[9377]: Accepted publickey for core from 139.178.89.65 port 35632 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:00.624558 sshd[9377]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:00.627086 systemd-logind[1809]: New session 19 of user core. Sep 13 01:28:00.637543 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 13 01:28:00.721477 sshd[9377]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:00.723007 systemd[1]: sshd@17-139.178.94.15:22-139.178.89.65:35632.service: Deactivated successfully. Sep 13 01:28:00.723937 systemd[1]: session-19.scope: Deactivated successfully. Sep 13 01:28:00.724654 systemd-logind[1809]: Session 19 logged out. Waiting for processes to exit. Sep 13 01:28:00.725198 systemd-logind[1809]: Removed session 19. Sep 13 01:28:05.774997 systemd[1]: Started sshd@18-139.178.94.15:22-139.178.89.65:35648.service - OpenSSH per-connection server daemon (139.178.89.65:35648). Sep 13 01:28:05.838622 sshd[9403]: Accepted publickey for core from 139.178.89.65 port 35648 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:05.839778 sshd[9403]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:05.843443 systemd-logind[1809]: New session 20 of user core. Sep 13 01:28:05.862508 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 13 01:28:06.000637 sshd[9403]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:06.020058 systemd[1]: sshd@18-139.178.94.15:22-139.178.89.65:35648.service: Deactivated successfully. Sep 13 01:28:06.020884 systemd[1]: session-20.scope: Deactivated successfully. Sep 13 01:28:06.021518 systemd-logind[1809]: Session 20 logged out. Waiting for processes to exit. Sep 13 01:28:06.022183 systemd[1]: Started sshd@19-139.178.94.15:22-139.178.89.65:35652.service - OpenSSH per-connection server daemon (139.178.89.65:35652). Sep 13 01:28:06.022625 systemd-logind[1809]: Removed session 20. Sep 13 01:28:06.094259 sshd[9429]: Accepted publickey for core from 139.178.89.65 port 35652 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:06.095675 sshd[9429]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:06.100014 systemd-logind[1809]: New session 21 of user core. Sep 13 01:28:06.109438 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 13 01:28:06.320188 sshd[9429]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:06.346921 systemd[1]: sshd@19-139.178.94.15:22-139.178.89.65:35652.service: Deactivated successfully. Sep 13 01:28:06.348440 systemd[1]: session-21.scope: Deactivated successfully. Sep 13 01:28:06.349799 systemd-logind[1809]: Session 21 logged out. Waiting for processes to exit. Sep 13 01:28:06.351123 systemd[1]: Started sshd@20-139.178.94.15:22-139.178.89.65:35666.service - OpenSSH per-connection server daemon (139.178.89.65:35666). Sep 13 01:28:06.352190 systemd-logind[1809]: Removed session 21. Sep 13 01:28:06.413499 sshd[9454]: Accepted publickey for core from 139.178.89.65 port 35666 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:06.414930 sshd[9454]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:06.419891 systemd-logind[1809]: New session 22 of user core. Sep 13 01:28:06.437450 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 13 01:28:07.077986 sshd[9454]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:07.089464 systemd[1]: sshd@20-139.178.94.15:22-139.178.89.65:35666.service: Deactivated successfully. Sep 13 01:28:07.090687 systemd[1]: session-22.scope: Deactivated successfully. Sep 13 01:28:07.091642 systemd-logind[1809]: Session 22 logged out. Waiting for processes to exit. Sep 13 01:28:07.092636 systemd[1]: Started sshd@21-139.178.94.15:22-139.178.89.65:35680.service - OpenSSH per-connection server daemon (139.178.89.65:35680). Sep 13 01:28:07.093256 systemd-logind[1809]: Removed session 22. Sep 13 01:28:07.134256 sshd[9520]: Accepted publickey for core from 139.178.89.65 port 35680 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:07.135280 sshd[9520]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:07.138880 systemd-logind[1809]: New session 23 of user core. Sep 13 01:28:07.149383 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 13 01:28:07.351315 sshd[9520]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:07.365083 systemd[1]: sshd@21-139.178.94.15:22-139.178.89.65:35680.service: Deactivated successfully. Sep 13 01:28:07.365950 systemd[1]: session-23.scope: Deactivated successfully. Sep 13 01:28:07.366652 systemd-logind[1809]: Session 23 logged out. Waiting for processes to exit. Sep 13 01:28:07.367255 systemd[1]: Started sshd@22-139.178.94.15:22-139.178.89.65:35694.service - OpenSSH per-connection server daemon (139.178.89.65:35694). Sep 13 01:28:07.367765 systemd-logind[1809]: Removed session 23. Sep 13 01:28:07.398164 sshd[9547]: Accepted publickey for core from 139.178.89.65 port 35694 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:07.398829 sshd[9547]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:07.401485 systemd-logind[1809]: New session 24 of user core. Sep 13 01:28:07.413391 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 13 01:28:07.539038 sshd[9547]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:07.540593 systemd[1]: sshd@22-139.178.94.15:22-139.178.89.65:35694.service: Deactivated successfully. Sep 13 01:28:07.541520 systemd[1]: session-24.scope: Deactivated successfully. Sep 13 01:28:07.542226 systemd-logind[1809]: Session 24 logged out. Waiting for processes to exit. Sep 13 01:28:07.542951 systemd-logind[1809]: Removed session 24. Sep 13 01:28:12.559532 systemd[1]: Started sshd@23-139.178.94.15:22-139.178.89.65:55754.service - OpenSSH per-connection server daemon (139.178.89.65:55754). Sep 13 01:28:12.593767 sshd[9577]: Accepted publickey for core from 139.178.89.65 port 55754 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:12.594432 sshd[9577]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:12.596751 systemd-logind[1809]: New session 25 of user core. Sep 13 01:28:12.616381 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 13 01:28:12.702204 sshd[9577]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:12.703992 systemd[1]: sshd@23-139.178.94.15:22-139.178.89.65:55754.service: Deactivated successfully. Sep 13 01:28:12.704975 systemd[1]: session-25.scope: Deactivated successfully. Sep 13 01:28:12.705764 systemd-logind[1809]: Session 25 logged out. Waiting for processes to exit. Sep 13 01:28:12.706282 systemd-logind[1809]: Removed session 25. Sep 13 01:28:17.720056 systemd[1]: Started sshd@24-139.178.94.15:22-139.178.89.65:55764.service - OpenSSH per-connection server daemon (139.178.89.65:55764). Sep 13 01:28:17.751099 sshd[9634]: Accepted publickey for core from 139.178.89.65 port 55764 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:17.751764 sshd[9634]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:17.754453 systemd-logind[1809]: New session 26 of user core. Sep 13 01:28:17.769417 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 13 01:28:17.854593 sshd[9634]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:17.856145 systemd[1]: sshd@24-139.178.94.15:22-139.178.89.65:55764.service: Deactivated successfully. Sep 13 01:28:17.857043 systemd[1]: session-26.scope: Deactivated successfully. Sep 13 01:28:17.857751 systemd-logind[1809]: Session 26 logged out. Waiting for processes to exit. Sep 13 01:28:17.858228 systemd-logind[1809]: Removed session 26. Sep 13 01:28:22.877100 systemd[1]: Started sshd@25-139.178.94.15:22-139.178.89.65:47764.service - OpenSSH per-connection server daemon (139.178.89.65:47764). Sep 13 01:28:22.949142 sshd[9710]: Accepted publickey for core from 139.178.89.65 port 47764 ssh2: RSA SHA256:9yt090AVdPEq/FQCZmOXJ9hsscYfxbTJezbW0JfpgHU Sep 13 01:28:22.951449 sshd[9710]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:28:22.957164 systemd-logind[1809]: New session 27 of user core. Sep 13 01:28:22.966542 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 13 01:28:23.049474 sshd[9710]: pam_unix(sshd:session): session closed for user core Sep 13 01:28:23.051147 systemd[1]: sshd@25-139.178.94.15:22-139.178.89.65:47764.service: Deactivated successfully. Sep 13 01:28:23.052123 systemd[1]: session-27.scope: Deactivated successfully. Sep 13 01:28:23.052835 systemd-logind[1809]: Session 27 logged out. Waiting for processes to exit. Sep 13 01:28:23.053446 systemd-logind[1809]: Removed session 27.