Nov 8 01:37:37.027764 kernel: Linux version 6.6.113-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Nov 7 22:45:04 -00 2025 Nov 8 01:37:37.027778 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=480a02cf7f2001774aa495c3e719d4173e968e6839485a7d2b207ef2facca472 Nov 8 01:37:37.027785 kernel: BIOS-provided physical RAM map: Nov 8 01:37:37.027789 kernel: BIOS-e820: [mem 0x0000000000000000-0x00000000000997ff] usable Nov 8 01:37:37.027793 kernel: BIOS-e820: [mem 0x0000000000099800-0x000000000009ffff] reserved Nov 8 01:37:37.027797 kernel: BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved Nov 8 01:37:37.027802 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003fffffff] usable Nov 8 01:37:37.027806 kernel: BIOS-e820: [mem 0x0000000040000000-0x00000000403fffff] reserved Nov 8 01:37:37.027810 kernel: BIOS-e820: [mem 0x0000000040400000-0x00000000825bdfff] usable Nov 8 01:37:37.027814 kernel: BIOS-e820: [mem 0x00000000825be000-0x00000000825befff] ACPI NVS Nov 8 01:37:37.027819 kernel: BIOS-e820: [mem 0x00000000825bf000-0x00000000825bffff] reserved Nov 8 01:37:37.027824 kernel: BIOS-e820: [mem 0x00000000825c0000-0x000000008afcdfff] usable Nov 8 01:37:37.027828 kernel: BIOS-e820: [mem 0x000000008afce000-0x000000008c0b2fff] reserved Nov 8 01:37:37.027832 kernel: BIOS-e820: [mem 0x000000008c0b3000-0x000000008c23bfff] usable Nov 8 01:37:37.027837 kernel: BIOS-e820: [mem 0x000000008c23c000-0x000000008c66dfff] ACPI NVS Nov 8 01:37:37.027842 kernel: BIOS-e820: [mem 0x000000008c66e000-0x000000008eefefff] reserved Nov 8 01:37:37.027848 kernel: BIOS-e820: [mem 0x000000008eeff000-0x000000008eefffff] usable Nov 8 01:37:37.027852 kernel: BIOS-e820: [mem 0x000000008ef00000-0x000000008fffffff] reserved Nov 8 01:37:37.027857 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Nov 8 01:37:37.027862 kernel: BIOS-e820: [mem 0x00000000fe000000-0x00000000fe010fff] reserved Nov 8 01:37:37.027867 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec00fff] reserved Nov 8 01:37:37.027871 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Nov 8 01:37:37.027876 kernel: BIOS-e820: [mem 0x00000000ff000000-0x00000000ffffffff] reserved Nov 8 01:37:37.027881 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000086effffff] usable Nov 8 01:37:37.027885 kernel: NX (Execute Disable) protection: active Nov 8 01:37:37.027890 kernel: APIC: Static calls initialized Nov 8 01:37:37.027895 kernel: SMBIOS 3.2.1 present. Nov 8 01:37:37.027900 kernel: DMI: Supermicro SYS-5019C-MR-PH004/X11SCM-F, BIOS 2.6 12/03/2024 Nov 8 01:37:37.027906 kernel: tsc: Detected 3400.000 MHz processor Nov 8 01:37:37.027910 kernel: tsc: Detected 3399.906 MHz TSC Nov 8 01:37:37.027915 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Nov 8 01:37:37.027921 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Nov 8 01:37:37.027926 kernel: last_pfn = 0x86f000 max_arch_pfn = 0x400000000 Nov 8 01:37:37.027930 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 23), built from 10 variable MTRRs Nov 8 01:37:37.027935 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Nov 8 01:37:37.027940 kernel: last_pfn = 0x8ef00 max_arch_pfn = 0x400000000 Nov 8 01:37:37.027945 kernel: Using GB pages for direct mapping Nov 8 01:37:37.027951 kernel: ACPI: Early table checksum verification disabled Nov 8 01:37:37.027955 kernel: ACPI: RSDP 0x00000000000F05B0 000024 (v02 SUPERM) Nov 8 01:37:37.027960 kernel: ACPI: XSDT 0x000000008C54F0C8 00010C (v01 SUPERM SUPERM 01072009 AMI 00010013) Nov 8 01:37:37.027967 kernel: ACPI: FACP 0x000000008C58B670 000114 (v06 01072009 AMI 00010013) Nov 8 01:37:37.027972 kernel: ACPI: DSDT 0x000000008C54F268 03C404 (v02 SUPERM SMCI--MB 01072009 INTL 20160527) Nov 8 01:37:37.027977 kernel: ACPI: FACS 0x000000008C66DF80 000040 Nov 8 01:37:37.027983 kernel: ACPI: APIC 0x000000008C58B788 00012C (v04 01072009 AMI 00010013) Nov 8 01:37:37.027989 kernel: ACPI: FPDT 0x000000008C58B8B8 000044 (v01 01072009 AMI 00010013) Nov 8 01:37:37.027994 kernel: ACPI: FIDT 0x000000008C58B900 00009C (v01 SUPERM SMCI--MB 01072009 AMI 00010013) Nov 8 01:37:37.027999 kernel: ACPI: MCFG 0x000000008C58B9A0 00003C (v01 SUPERM SMCI--MB 01072009 MSFT 00000097) Nov 8 01:37:37.028004 kernel: ACPI: SPMI 0x000000008C58B9E0 000041 (v05 SUPERM SMCI--MB 00000000 AMI. 00000000) Nov 8 01:37:37.028009 kernel: ACPI: SSDT 0x000000008C58BA28 001B1C (v02 CpuRef CpuSsdt 00003000 INTL 20160527) Nov 8 01:37:37.028014 kernel: ACPI: SSDT 0x000000008C58D548 0031C6 (v02 SaSsdt SaSsdt 00003000 INTL 20160527) Nov 8 01:37:37.028019 kernel: ACPI: SSDT 0x000000008C590710 00232B (v02 PegSsd PegSsdt 00001000 INTL 20160527) Nov 8 01:37:37.028025 kernel: ACPI: HPET 0x000000008C592A40 000038 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028030 kernel: ACPI: SSDT 0x000000008C592A78 000FAE (v02 SUPERM Ther_Rvp 00001000 INTL 20160527) Nov 8 01:37:37.028035 kernel: ACPI: SSDT 0x000000008C593A28 0008F4 (v02 INTEL xh_mossb 00000000 INTL 20160527) Nov 8 01:37:37.028040 kernel: ACPI: UEFI 0x000000008C594320 000042 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028045 kernel: ACPI: LPIT 0x000000008C594368 000094 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028050 kernel: ACPI: SSDT 0x000000008C594400 0027DE (v02 SUPERM PtidDevc 00001000 INTL 20160527) Nov 8 01:37:37.028056 kernel: ACPI: SSDT 0x000000008C596BE0 0014E2 (v02 SUPERM TbtTypeC 00000000 INTL 20160527) Nov 8 01:37:37.028061 kernel: ACPI: DBGP 0x000000008C5980C8 000034 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028066 kernel: ACPI: DBG2 0x000000008C598100 000054 (v00 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028072 kernel: ACPI: SSDT 0x000000008C598158 001B67 (v02 SUPERM UsbCTabl 00001000 INTL 20160527) Nov 8 01:37:37.028077 kernel: ACPI: DMAR 0x000000008C599CC0 000070 (v01 INTEL EDK2 00000002 01000013) Nov 8 01:37:37.028082 kernel: ACPI: SSDT 0x000000008C599D30 000144 (v02 Intel ADebTabl 00001000 INTL 20160527) Nov 8 01:37:37.028087 kernel: ACPI: TPM2 0x000000008C599E78 000034 (v04 SUPERM SMCI--MB 00000001 AMI 00000000) Nov 8 01:37:37.028092 kernel: ACPI: SSDT 0x000000008C599EB0 000D8F (v02 INTEL SpsNm 00000002 INTL 20160527) Nov 8 01:37:37.028097 kernel: ACPI: WSMT 0x000000008C59AC40 000028 (v01 SUPERM 01072009 AMI 00010013) Nov 8 01:37:37.028102 kernel: ACPI: EINJ 0x000000008C59AC68 000130 (v01 AMI AMI.EINJ 00000000 AMI. 00000000) Nov 8 01:37:37.028107 kernel: ACPI: ERST 0x000000008C59AD98 000230 (v01 AMIER AMI.ERST 00000000 AMI. 00000000) Nov 8 01:37:37.028113 kernel: ACPI: BERT 0x000000008C59AFC8 000030 (v01 AMI AMI.BERT 00000000 AMI. 00000000) Nov 8 01:37:37.028118 kernel: ACPI: HEST 0x000000008C59AFF8 00027C (v01 AMI AMI.HEST 00000000 AMI. 00000000) Nov 8 01:37:37.028123 kernel: ACPI: SSDT 0x000000008C59B278 000162 (v01 SUPERM SMCCDN 00000000 INTL 20181221) Nov 8 01:37:37.028128 kernel: ACPI: Reserving FACP table memory at [mem 0x8c58b670-0x8c58b783] Nov 8 01:37:37.028133 kernel: ACPI: Reserving DSDT table memory at [mem 0x8c54f268-0x8c58b66b] Nov 8 01:37:37.028138 kernel: ACPI: Reserving FACS table memory at [mem 0x8c66df80-0x8c66dfbf] Nov 8 01:37:37.028147 kernel: ACPI: Reserving APIC table memory at [mem 0x8c58b788-0x8c58b8b3] Nov 8 01:37:37.028152 kernel: ACPI: Reserving FPDT table memory at [mem 0x8c58b8b8-0x8c58b8fb] Nov 8 01:37:37.028177 kernel: ACPI: Reserving FIDT table memory at [mem 0x8c58b900-0x8c58b99b] Nov 8 01:37:37.028183 kernel: ACPI: Reserving MCFG table memory at [mem 0x8c58b9a0-0x8c58b9db] Nov 8 01:37:37.028189 kernel: ACPI: Reserving SPMI table memory at [mem 0x8c58b9e0-0x8c58ba20] Nov 8 01:37:37.028194 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58ba28-0x8c58d543] Nov 8 01:37:37.028212 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58d548-0x8c59070d] Nov 8 01:37:37.028217 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c590710-0x8c592a3a] Nov 8 01:37:37.028222 kernel: ACPI: Reserving HPET table memory at [mem 0x8c592a40-0x8c592a77] Nov 8 01:37:37.028227 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c592a78-0x8c593a25] Nov 8 01:37:37.028232 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c593a28-0x8c59431b] Nov 8 01:37:37.028237 kernel: ACPI: Reserving UEFI table memory at [mem 0x8c594320-0x8c594361] Nov 8 01:37:37.028243 kernel: ACPI: Reserving LPIT table memory at [mem 0x8c594368-0x8c5943fb] Nov 8 01:37:37.028248 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c594400-0x8c596bdd] Nov 8 01:37:37.028253 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c596be0-0x8c5980c1] Nov 8 01:37:37.028258 kernel: ACPI: Reserving DBGP table memory at [mem 0x8c5980c8-0x8c5980fb] Nov 8 01:37:37.028263 kernel: ACPI: Reserving DBG2 table memory at [mem 0x8c598100-0x8c598153] Nov 8 01:37:37.028268 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c598158-0x8c599cbe] Nov 8 01:37:37.028273 kernel: ACPI: Reserving DMAR table memory at [mem 0x8c599cc0-0x8c599d2f] Nov 8 01:37:37.028278 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599d30-0x8c599e73] Nov 8 01:37:37.028283 kernel: ACPI: Reserving TPM2 table memory at [mem 0x8c599e78-0x8c599eab] Nov 8 01:37:37.028289 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599eb0-0x8c59ac3e] Nov 8 01:37:37.028294 kernel: ACPI: Reserving WSMT table memory at [mem 0x8c59ac40-0x8c59ac67] Nov 8 01:37:37.028299 kernel: ACPI: Reserving EINJ table memory at [mem 0x8c59ac68-0x8c59ad97] Nov 8 01:37:37.028304 kernel: ACPI: Reserving ERST table memory at [mem 0x8c59ad98-0x8c59afc7] Nov 8 01:37:37.028309 kernel: ACPI: Reserving BERT table memory at [mem 0x8c59afc8-0x8c59aff7] Nov 8 01:37:37.028315 kernel: ACPI: Reserving HEST table memory at [mem 0x8c59aff8-0x8c59b273] Nov 8 01:37:37.028320 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c59b278-0x8c59b3d9] Nov 8 01:37:37.028325 kernel: No NUMA configuration found Nov 8 01:37:37.028330 kernel: Faking a node at [mem 0x0000000000000000-0x000000086effffff] Nov 8 01:37:37.028335 kernel: NODE_DATA(0) allocated [mem 0x86effa000-0x86effffff] Nov 8 01:37:37.028341 kernel: Zone ranges: Nov 8 01:37:37.028346 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Nov 8 01:37:37.028351 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Nov 8 01:37:37.028356 kernel: Normal [mem 0x0000000100000000-0x000000086effffff] Nov 8 01:37:37.028362 kernel: Movable zone start for each node Nov 8 01:37:37.028367 kernel: Early memory node ranges Nov 8 01:37:37.028372 kernel: node 0: [mem 0x0000000000001000-0x0000000000098fff] Nov 8 01:37:37.028377 kernel: node 0: [mem 0x0000000000100000-0x000000003fffffff] Nov 8 01:37:37.028382 kernel: node 0: [mem 0x0000000040400000-0x00000000825bdfff] Nov 8 01:37:37.028388 kernel: node 0: [mem 0x00000000825c0000-0x000000008afcdfff] Nov 8 01:37:37.028393 kernel: node 0: [mem 0x000000008c0b3000-0x000000008c23bfff] Nov 8 01:37:37.028398 kernel: node 0: [mem 0x000000008eeff000-0x000000008eefffff] Nov 8 01:37:37.028403 kernel: node 0: [mem 0x0000000100000000-0x000000086effffff] Nov 8 01:37:37.028412 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000086effffff] Nov 8 01:37:37.028418 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Nov 8 01:37:37.028424 kernel: On node 0, zone DMA: 103 pages in unavailable ranges Nov 8 01:37:37.028429 kernel: On node 0, zone DMA32: 1024 pages in unavailable ranges Nov 8 01:37:37.028435 kernel: On node 0, zone DMA32: 2 pages in unavailable ranges Nov 8 01:37:37.028441 kernel: On node 0, zone DMA32: 4325 pages in unavailable ranges Nov 8 01:37:37.028446 kernel: On node 0, zone DMA32: 11459 pages in unavailable ranges Nov 8 01:37:37.028452 kernel: On node 0, zone Normal: 4352 pages in unavailable ranges Nov 8 01:37:37.028458 kernel: On node 0, zone Normal: 4096 pages in unavailable ranges Nov 8 01:37:37.028463 kernel: ACPI: PM-Timer IO Port: 0x1808 Nov 8 01:37:37.028468 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Nov 8 01:37:37.028474 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Nov 8 01:37:37.028479 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Nov 8 01:37:37.028486 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Nov 8 01:37:37.028491 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Nov 8 01:37:37.028496 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Nov 8 01:37:37.028502 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Nov 8 01:37:37.028507 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Nov 8 01:37:37.028513 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Nov 8 01:37:37.028518 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Nov 8 01:37:37.028523 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Nov 8 01:37:37.028529 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Nov 8 01:37:37.028535 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Nov 8 01:37:37.028540 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Nov 8 01:37:37.028546 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Nov 8 01:37:37.028551 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Nov 8 01:37:37.028556 kernel: IOAPIC[0]: apic_id 2, version 32, address 0xfec00000, GSI 0-119 Nov 8 01:37:37.028562 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Nov 8 01:37:37.028567 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Nov 8 01:37:37.028573 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Nov 8 01:37:37.028578 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Nov 8 01:37:37.028585 kernel: TSC deadline timer available Nov 8 01:37:37.028590 kernel: smpboot: Allowing 16 CPUs, 0 hotplug CPUs Nov 8 01:37:37.028596 kernel: [mem 0x90000000-0xdfffffff] available for PCI devices Nov 8 01:37:37.028601 kernel: Booting paravirtualized kernel on bare hardware Nov 8 01:37:37.028607 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Nov 8 01:37:37.028612 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Nov 8 01:37:37.028618 kernel: percpu: Embedded 58 pages/cpu s196712 r8192 d32664 u262144 Nov 8 01:37:37.028623 kernel: pcpu-alloc: s196712 r8192 d32664 u262144 alloc=1*2097152 Nov 8 01:37:37.028628 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Nov 8 01:37:37.028635 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=480a02cf7f2001774aa495c3e719d4173e968e6839485a7d2b207ef2facca472 Nov 8 01:37:37.028641 kernel: random: crng init done Nov 8 01:37:37.028646 kernel: Dentry cache hash table entries: 4194304 (order: 13, 33554432 bytes, linear) Nov 8 01:37:37.028652 kernel: Inode-cache hash table entries: 2097152 (order: 12, 16777216 bytes, linear) Nov 8 01:37:37.028657 kernel: Fallback order for Node 0: 0 Nov 8 01:37:37.028663 kernel: Built 1 zonelists, mobility grouping on. Total pages: 8232416 Nov 8 01:37:37.028668 kernel: Policy zone: Normal Nov 8 01:37:37.028674 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 8 01:37:37.028680 kernel: software IO TLB: area num 16. Nov 8 01:37:37.028686 kernel: Memory: 32720312K/33452984K available (12288K kernel code, 2288K rwdata, 22748K rodata, 42880K init, 2320K bss, 732412K reserved, 0K cma-reserved) Nov 8 01:37:37.028692 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Nov 8 01:37:37.028697 kernel: ftrace: allocating 37980 entries in 149 pages Nov 8 01:37:37.028702 kernel: ftrace: allocated 149 pages with 4 groups Nov 8 01:37:37.028708 kernel: Dynamic Preempt: voluntary Nov 8 01:37:37.028714 kernel: rcu: Preemptible hierarchical RCU implementation. Nov 8 01:37:37.028719 kernel: rcu: RCU event tracing is enabled. Nov 8 01:37:37.028725 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Nov 8 01:37:37.028732 kernel: Trampoline variant of Tasks RCU enabled. Nov 8 01:37:37.028738 kernel: Rude variant of Tasks RCU enabled. Nov 8 01:37:37.028743 kernel: Tracing variant of Tasks RCU enabled. Nov 8 01:37:37.028748 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 8 01:37:37.028754 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Nov 8 01:37:37.028759 kernel: NR_IRQS: 33024, nr_irqs: 2184, preallocated irqs: 16 Nov 8 01:37:37.028765 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 8 01:37:37.028770 kernel: Console: colour dummy device 80x25 Nov 8 01:37:37.028776 kernel: printk: console [tty0] enabled Nov 8 01:37:37.028781 kernel: printk: console [ttyS1] enabled Nov 8 01:37:37.028788 kernel: ACPI: Core revision 20230628 Nov 8 01:37:37.028793 kernel: hpet: HPET dysfunctional in PC10. Force disabled. Nov 8 01:37:37.028799 kernel: APIC: Switch to symmetric I/O mode setup Nov 8 01:37:37.028804 kernel: DMAR: Host address width 39 Nov 8 01:37:37.028810 kernel: DMAR: DRHD base: 0x000000fed91000 flags: 0x1 Nov 8 01:37:37.028815 kernel: DMAR: dmar0: reg_base_addr fed91000 ver 1:0 cap d2008c40660462 ecap f050da Nov 8 01:37:37.028821 kernel: DMAR: RMRR base: 0x0000008cf19000 end: 0x0000008d162fff Nov 8 01:37:37.028826 kernel: DMAR-IR: IOAPIC id 2 under DRHD base 0xfed91000 IOMMU 0 Nov 8 01:37:37.028832 kernel: DMAR-IR: HPET id 0 under DRHD base 0xfed91000 Nov 8 01:37:37.028838 kernel: DMAR-IR: Queued invalidation will be enabled to support x2apic and Intr-remapping. Nov 8 01:37:37.028843 kernel: DMAR-IR: Enabled IRQ remapping in x2apic mode Nov 8 01:37:37.028849 kernel: x2apic enabled Nov 8 01:37:37.028855 kernel: APIC: Switched APIC routing to: cluster x2apic Nov 8 01:37:37.028860 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x3101f59f5e6, max_idle_ns: 440795259996 ns Nov 8 01:37:37.028866 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 6799.81 BogoMIPS (lpj=3399906) Nov 8 01:37:37.028871 kernel: CPU0: Thermal monitoring enabled (TM1) Nov 8 01:37:37.028877 kernel: process: using mwait in idle threads Nov 8 01:37:37.028882 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Nov 8 01:37:37.028889 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Nov 8 01:37:37.028894 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Nov 8 01:37:37.028899 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Nov 8 01:37:37.028905 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Nov 8 01:37:37.028910 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Nov 8 01:37:37.028915 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Nov 8 01:37:37.028921 kernel: RETBleed: Mitigation: Enhanced IBRS Nov 8 01:37:37.028926 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Nov 8 01:37:37.028931 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Nov 8 01:37:37.028937 kernel: TAA: Mitigation: TSX disabled Nov 8 01:37:37.028942 kernel: MMIO Stale Data: Mitigation: Clear CPU buffers Nov 8 01:37:37.028948 kernel: SRBDS: Mitigation: Microcode Nov 8 01:37:37.028954 kernel: GDS: Mitigation: Microcode Nov 8 01:37:37.028959 kernel: active return thunk: its_return_thunk Nov 8 01:37:37.028965 kernel: ITS: Mitigation: Aligned branch/return thunks Nov 8 01:37:37.028970 kernel: VMSCAPE: Mitigation: IBPB before exit to userspace Nov 8 01:37:37.028975 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Nov 8 01:37:37.028981 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Nov 8 01:37:37.028986 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Nov 8 01:37:37.028992 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Nov 8 01:37:37.028997 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Nov 8 01:37:37.029002 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Nov 8 01:37:37.029009 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Nov 8 01:37:37.029014 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Nov 8 01:37:37.029019 kernel: x86/fpu: Enabled xstate features 0x1f, context size is 960 bytes, using 'compacted' format. Nov 8 01:37:37.029025 kernel: Freeing SMP alternatives memory: 32K Nov 8 01:37:37.029030 kernel: pid_max: default: 32768 minimum: 301 Nov 8 01:37:37.029036 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Nov 8 01:37:37.029041 kernel: landlock: Up and running. Nov 8 01:37:37.029046 kernel: SELinux: Initializing. Nov 8 01:37:37.029052 kernel: Mount-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.029057 kernel: Mountpoint-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.029063 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Nov 8 01:37:37.029068 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Nov 8 01:37:37.029074 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Nov 8 01:37:37.029080 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Nov 8 01:37:37.029086 kernel: Performance Events: PEBS fmt3+, Skylake events, 32-deep LBR, full-width counters, Intel PMU driver. Nov 8 01:37:37.029091 kernel: ... version: 4 Nov 8 01:37:37.029097 kernel: ... bit width: 48 Nov 8 01:37:37.029102 kernel: ... generic registers: 4 Nov 8 01:37:37.029107 kernel: ... value mask: 0000ffffffffffff Nov 8 01:37:37.029113 kernel: ... max period: 00007fffffffffff Nov 8 01:37:37.029118 kernel: ... fixed-purpose events: 3 Nov 8 01:37:37.029125 kernel: ... event mask: 000000070000000f Nov 8 01:37:37.029130 kernel: signal: max sigframe size: 2032 Nov 8 01:37:37.029136 kernel: Estimated ratio of average max frequency by base frequency (times 1024): 1445 Nov 8 01:37:37.029147 kernel: rcu: Hierarchical SRCU implementation. Nov 8 01:37:37.029153 kernel: rcu: Max phase no-delay instances is 400. Nov 8 01:37:37.029179 kernel: NMI watchdog: Enabled. Permanently consumes one hw-PMU counter. Nov 8 01:37:37.029185 kernel: smp: Bringing up secondary CPUs ... Nov 8 01:37:37.029191 kernel: smpboot: x86: Booting SMP configuration: Nov 8 01:37:37.029211 kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6 #7 #8 #9 #10 #11 #12 #13 #14 #15 Nov 8 01:37:37.029218 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Nov 8 01:37:37.029224 kernel: smp: Brought up 1 node, 16 CPUs Nov 8 01:37:37.029229 kernel: smpboot: Max logical packages: 1 Nov 8 01:37:37.029234 kernel: smpboot: Total of 16 processors activated (108796.99 BogoMIPS) Nov 8 01:37:37.029240 kernel: devtmpfs: initialized Nov 8 01:37:37.029245 kernel: x86/mm: Memory block size: 128MB Nov 8 01:37:37.029251 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x825be000-0x825befff] (4096 bytes) Nov 8 01:37:37.029256 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x8c23c000-0x8c66dfff] (4399104 bytes) Nov 8 01:37:37.029263 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 8 01:37:37.029268 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Nov 8 01:37:37.029274 kernel: pinctrl core: initialized pinctrl subsystem Nov 8 01:37:37.029279 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 8 01:37:37.029284 kernel: audit: initializing netlink subsys (disabled) Nov 8 01:37:37.029290 kernel: audit: type=2000 audit(1762565851.039:1): state=initialized audit_enabled=0 res=1 Nov 8 01:37:37.029295 kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 8 01:37:37.029301 kernel: thermal_sys: Registered thermal governor 'user_space' Nov 8 01:37:37.029306 kernel: cpuidle: using governor menu Nov 8 01:37:37.029313 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 8 01:37:37.029318 kernel: dca service started, version 1.12.1 Nov 8 01:37:37.029324 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Nov 8 01:37:37.029329 kernel: PCI: Using configuration type 1 for base access Nov 8 01:37:37.029335 kernel: ENERGY_PERF_BIAS: Set to 'normal', was 'performance' Nov 8 01:37:37.029340 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Nov 8 01:37:37.029345 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Nov 8 01:37:37.029351 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Nov 8 01:37:37.029356 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Nov 8 01:37:37.029363 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Nov 8 01:37:37.029368 kernel: ACPI: Added _OSI(Module Device) Nov 8 01:37:37.029374 kernel: ACPI: Added _OSI(Processor Device) Nov 8 01:37:37.029379 kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 8 01:37:37.029384 kernel: ACPI: 12 ACPI AML tables successfully acquired and loaded Nov 8 01:37:37.029390 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029395 kernel: ACPI: SSDT 0xFFFF9E9CC1B36800 000400 (v02 PmRef Cpu0Cst 00003001 INTL 20160527) Nov 8 01:37:37.029401 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029406 kernel: ACPI: SSDT 0xFFFF9E9CC1B2A000 000683 (v02 PmRef Cpu0Ist 00003000 INTL 20160527) Nov 8 01:37:37.029413 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029418 kernel: ACPI: SSDT 0xFFFF9E9CC0247400 0000F4 (v02 PmRef Cpu0Psd 00003000 INTL 20160527) Nov 8 01:37:37.029424 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029429 kernel: ACPI: SSDT 0xFFFF9E9CC1E5C800 0005FC (v02 PmRef ApIst 00003000 INTL 20160527) Nov 8 01:37:37.029434 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029439 kernel: ACPI: SSDT 0xFFFF9E9CC012F000 000AB0 (v02 PmRef ApPsd 00003000 INTL 20160527) Nov 8 01:37:37.029445 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029450 kernel: ACPI: SSDT 0xFFFF9E9CC1B30400 00030A (v02 PmRef ApCst 00003000 INTL 20160527) Nov 8 01:37:37.029456 kernel: ACPI: _OSC evaluated successfully for all CPUs Nov 8 01:37:37.029461 kernel: ACPI: Interpreter enabled Nov 8 01:37:37.029468 kernel: ACPI: PM: (supports S0 S5) Nov 8 01:37:37.029473 kernel: ACPI: Using IOAPIC for interrupt routing Nov 8 01:37:37.029478 kernel: HEST: Enabling Firmware First mode for corrected errors. Nov 8 01:37:37.029484 kernel: mce: [Firmware Bug]: Ignoring request to disable invalid MCA bank 14. Nov 8 01:37:37.029489 kernel: HEST: Table parsing has been initialized. Nov 8 01:37:37.029495 kernel: GHES: APEI firmware first mode is enabled by APEI bit and WHEA _OSC. Nov 8 01:37:37.029500 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Nov 8 01:37:37.029506 kernel: PCI: Ignoring E820 reservations for host bridge windows Nov 8 01:37:37.029511 kernel: ACPI: Enabled 9 GPEs in block 00 to 7F Nov 8 01:37:37.029518 kernel: ACPI: \_SB_.PCI0.XDCI.USBC: New power resource Nov 8 01:37:37.029523 kernel: ACPI: \_SB_.PCI0.SAT0.VOL0.V0PR: New power resource Nov 8 01:37:37.029529 kernel: ACPI: \_SB_.PCI0.SAT0.VOL1.V1PR: New power resource Nov 8 01:37:37.029534 kernel: ACPI: \_SB_.PCI0.SAT0.VOL2.V2PR: New power resource Nov 8 01:37:37.029540 kernel: ACPI: \_SB_.PCI0.CNVW.WRST: New power resource Nov 8 01:37:37.029545 kernel: ACPI: \_TZ_.FN00: New power resource Nov 8 01:37:37.029551 kernel: ACPI: \_TZ_.FN01: New power resource Nov 8 01:37:37.029556 kernel: ACPI: \_TZ_.FN02: New power resource Nov 8 01:37:37.029561 kernel: ACPI: \_TZ_.FN03: New power resource Nov 8 01:37:37.029568 kernel: ACPI: \_TZ_.FN04: New power resource Nov 8 01:37:37.029573 kernel: ACPI: \PIN_: New power resource Nov 8 01:37:37.029579 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-fe]) Nov 8 01:37:37.029654 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Nov 8 01:37:37.029708 kernel: acpi PNP0A08:00: _OSC: platform does not support [AER] Nov 8 01:37:37.029758 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability LTR] Nov 8 01:37:37.029766 kernel: PCI host bridge to bus 0000:00 Nov 8 01:37:37.029817 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Nov 8 01:37:37.029862 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Nov 8 01:37:37.029905 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Nov 8 01:37:37.029948 kernel: pci_bus 0000:00: root bus resource [mem 0x90000000-0xdfffffff window] Nov 8 01:37:37.029989 kernel: pci_bus 0000:00: root bus resource [mem 0xfc800000-0xfe7fffff window] Nov 8 01:37:37.030032 kernel: pci_bus 0000:00: root bus resource [bus 00-fe] Nov 8 01:37:37.030090 kernel: pci 0000:00:00.0: [8086:3e31] type 00 class 0x060000 Nov 8 01:37:37.030153 kernel: pci 0000:00:01.0: [8086:1901] type 01 class 0x060400 Nov 8 01:37:37.030239 kernel: pci 0000:00:01.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.030291 kernel: pci 0000:00:08.0: [8086:1911] type 00 class 0x088000 Nov 8 01:37:37.030341 kernel: pci 0000:00:08.0: reg 0x10: [mem 0x9551f000-0x9551ffff 64bit] Nov 8 01:37:37.030393 kernel: pci 0000:00:12.0: [8086:a379] type 00 class 0x118000 Nov 8 01:37:37.030442 kernel: pci 0000:00:12.0: reg 0x10: [mem 0x9551e000-0x9551efff 64bit] Nov 8 01:37:37.030497 kernel: pci 0000:00:14.0: [8086:a36d] type 00 class 0x0c0330 Nov 8 01:37:37.030546 kernel: pci 0000:00:14.0: reg 0x10: [mem 0x95500000-0x9550ffff 64bit] Nov 8 01:37:37.030593 kernel: pci 0000:00:14.0: PME# supported from D3hot D3cold Nov 8 01:37:37.030646 kernel: pci 0000:00:14.2: [8086:a36f] type 00 class 0x050000 Nov 8 01:37:37.030695 kernel: pci 0000:00:14.2: reg 0x10: [mem 0x95512000-0x95513fff 64bit] Nov 8 01:37:37.030744 kernel: pci 0000:00:14.2: reg 0x18: [mem 0x9551d000-0x9551dfff 64bit] Nov 8 01:37:37.030798 kernel: pci 0000:00:15.0: [8086:a368] type 00 class 0x0c8000 Nov 8 01:37:37.030847 kernel: pci 0000:00:15.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Nov 8 01:37:37.030902 kernel: pci 0000:00:15.1: [8086:a369] type 00 class 0x0c8000 Nov 8 01:37:37.030951 kernel: pci 0000:00:15.1: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Nov 8 01:37:37.031003 kernel: pci 0000:00:16.0: [8086:a360] type 00 class 0x078000 Nov 8 01:37:37.031051 kernel: pci 0000:00:16.0: reg 0x10: [mem 0x9551a000-0x9551afff 64bit] Nov 8 01:37:37.031102 kernel: pci 0000:00:16.0: PME# supported from D3hot Nov 8 01:37:37.031174 kernel: pci 0000:00:16.1: [8086:a361] type 00 class 0x078000 Nov 8 01:37:37.031246 kernel: pci 0000:00:16.1: reg 0x10: [mem 0x95519000-0x95519fff 64bit] Nov 8 01:37:37.031295 kernel: pci 0000:00:16.1: PME# supported from D3hot Nov 8 01:37:37.031348 kernel: pci 0000:00:16.4: [8086:a364] type 00 class 0x078000 Nov 8 01:37:37.031396 kernel: pci 0000:00:16.4: reg 0x10: [mem 0x95518000-0x95518fff 64bit] Nov 8 01:37:37.031445 kernel: pci 0000:00:16.4: PME# supported from D3hot Nov 8 01:37:37.031500 kernel: pci 0000:00:17.0: [8086:a352] type 00 class 0x010601 Nov 8 01:37:37.031551 kernel: pci 0000:00:17.0: reg 0x10: [mem 0x95510000-0x95511fff] Nov 8 01:37:37.031599 kernel: pci 0000:00:17.0: reg 0x14: [mem 0x95517000-0x955170ff] Nov 8 01:37:37.031648 kernel: pci 0000:00:17.0: reg 0x18: [io 0x6050-0x6057] Nov 8 01:37:37.031695 kernel: pci 0000:00:17.0: reg 0x1c: [io 0x6040-0x6043] Nov 8 01:37:37.031743 kernel: pci 0000:00:17.0: reg 0x20: [io 0x6020-0x603f] Nov 8 01:37:37.031794 kernel: pci 0000:00:17.0: reg 0x24: [mem 0x95516000-0x955167ff] Nov 8 01:37:37.031842 kernel: pci 0000:00:17.0: PME# supported from D3hot Nov 8 01:37:37.031895 kernel: pci 0000:00:1b.0: [8086:a340] type 01 class 0x060400 Nov 8 01:37:37.031945 kernel: pci 0000:00:1b.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032001 kernel: pci 0000:00:1b.4: [8086:a32c] type 01 class 0x060400 Nov 8 01:37:37.032053 kernel: pci 0000:00:1b.4: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032106 kernel: pci 0000:00:1b.5: [8086:a32d] type 01 class 0x060400 Nov 8 01:37:37.032174 kernel: pci 0000:00:1b.5: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032242 kernel: pci 0000:00:1c.0: [8086:a338] type 01 class 0x060400 Nov 8 01:37:37.032291 kernel: pci 0000:00:1c.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032345 kernel: pci 0000:00:1c.3: [8086:a33b] type 01 class 0x060400 Nov 8 01:37:37.032397 kernel: pci 0000:00:1c.3: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032451 kernel: pci 0000:00:1e.0: [8086:a328] type 00 class 0x078000 Nov 8 01:37:37.032501 kernel: pci 0000:00:1e.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Nov 8 01:37:37.032554 kernel: pci 0000:00:1f.0: [8086:a309] type 00 class 0x060100 Nov 8 01:37:37.032608 kernel: pci 0000:00:1f.4: [8086:a323] type 00 class 0x0c0500 Nov 8 01:37:37.032658 kernel: pci 0000:00:1f.4: reg 0x10: [mem 0x95514000-0x955140ff 64bit] Nov 8 01:37:37.032709 kernel: pci 0000:00:1f.4: reg 0x20: [io 0xefa0-0xefbf] Nov 8 01:37:37.032761 kernel: pci 0000:00:1f.5: [8086:a324] type 00 class 0x0c8000 Nov 8 01:37:37.032811 kernel: pci 0000:00:1f.5: reg 0x10: [mem 0xfe010000-0xfe010fff] Nov 8 01:37:37.032865 kernel: pci 0000:01:00.0: [15b3:1015] type 00 class 0x020000 Nov 8 01:37:37.032917 kernel: pci 0000:01:00.0: reg 0x10: [mem 0x92000000-0x93ffffff 64bit pref] Nov 8 01:37:37.032968 kernel: pci 0000:01:00.0: reg 0x30: [mem 0x95200000-0x952fffff pref] Nov 8 01:37:37.033021 kernel: pci 0000:01:00.0: PME# supported from D3cold Nov 8 01:37:37.033071 kernel: pci 0000:01:00.0: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Nov 8 01:37:37.033122 kernel: pci 0000:01:00.0: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Nov 8 01:37:37.033214 kernel: pci 0000:01:00.1: [15b3:1015] type 00 class 0x020000 Nov 8 01:37:37.033265 kernel: pci 0000:01:00.1: reg 0x10: [mem 0x90000000-0x91ffffff 64bit pref] Nov 8 01:37:37.033317 kernel: pci 0000:01:00.1: reg 0x30: [mem 0x95100000-0x951fffff pref] Nov 8 01:37:37.033367 kernel: pci 0000:01:00.1: PME# supported from D3cold Nov 8 01:37:37.033420 kernel: pci 0000:01:00.1: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Nov 8 01:37:37.033470 kernel: pci 0000:01:00.1: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Nov 8 01:37:37.033520 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Nov 8 01:37:37.033569 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Nov 8 01:37:37.033620 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Nov 8 01:37:37.033670 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Nov 8 01:37:37.033725 kernel: pci 0000:03:00.0: working around ROM BAR overlap defect Nov 8 01:37:37.033777 kernel: pci 0000:03:00.0: [8086:1533] type 00 class 0x020000 Nov 8 01:37:37.033829 kernel: pci 0000:03:00.0: reg 0x10: [mem 0x95400000-0x9547ffff] Nov 8 01:37:37.033880 kernel: pci 0000:03:00.0: reg 0x18: [io 0x5000-0x501f] Nov 8 01:37:37.033930 kernel: pci 0000:03:00.0: reg 0x1c: [mem 0x95480000-0x95483fff] Nov 8 01:37:37.033980 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.034029 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Nov 8 01:37:37.034079 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Nov 8 01:37:37.034127 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Nov 8 01:37:37.034224 kernel: pci 0000:04:00.0: working around ROM BAR overlap defect Nov 8 01:37:37.034277 kernel: pci 0000:04:00.0: [8086:1533] type 00 class 0x020000 Nov 8 01:37:37.034327 kernel: pci 0000:04:00.0: reg 0x10: [mem 0x95300000-0x9537ffff] Nov 8 01:37:37.034378 kernel: pci 0000:04:00.0: reg 0x18: [io 0x4000-0x401f] Nov 8 01:37:37.034427 kernel: pci 0000:04:00.0: reg 0x1c: [mem 0x95380000-0x95383fff] Nov 8 01:37:37.034478 kernel: pci 0000:04:00.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.034527 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Nov 8 01:37:37.034576 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Nov 8 01:37:37.034628 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Nov 8 01:37:37.034678 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Nov 8 01:37:37.034733 kernel: pci 0000:06:00.0: [1a03:1150] type 01 class 0x060400 Nov 8 01:37:37.034787 kernel: pci 0000:06:00.0: enabling Extended Tags Nov 8 01:37:37.034837 kernel: pci 0000:06:00.0: supports D1 D2 Nov 8 01:37:37.034887 kernel: pci 0000:06:00.0: PME# supported from D0 D1 D2 D3hot D3cold Nov 8 01:37:37.034937 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Nov 8 01:37:37.034988 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.035039 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.035095 kernel: pci_bus 0000:07: extended config space not accessible Nov 8 01:37:37.035155 kernel: pci 0000:07:00.0: [1a03:2000] type 00 class 0x030000 Nov 8 01:37:37.035255 kernel: pci 0000:07:00.0: reg 0x10: [mem 0x94000000-0x94ffffff] Nov 8 01:37:37.035307 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x95000000-0x9501ffff] Nov 8 01:37:37.035359 kernel: pci 0000:07:00.0: reg 0x18: [io 0x3000-0x307f] Nov 8 01:37:37.035413 kernel: pci 0000:07:00.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Nov 8 01:37:37.035465 kernel: pci 0000:07:00.0: supports D1 D2 Nov 8 01:37:37.035518 kernel: pci 0000:07:00.0: PME# supported from D0 D1 D2 D3hot D3cold Nov 8 01:37:37.035568 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Nov 8 01:37:37.035618 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.035667 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.035675 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 0 Nov 8 01:37:37.035681 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 1 Nov 8 01:37:37.035689 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 0 Nov 8 01:37:37.035694 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 0 Nov 8 01:37:37.035700 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 0 Nov 8 01:37:37.035706 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 0 Nov 8 01:37:37.035712 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 0 Nov 8 01:37:37.035717 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 0 Nov 8 01:37:37.035723 kernel: iommu: Default domain type: Translated Nov 8 01:37:37.035729 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Nov 8 01:37:37.035735 kernel: PCI: Using ACPI for IRQ routing Nov 8 01:37:37.035741 kernel: PCI: pci_cache_line_size set to 64 bytes Nov 8 01:37:37.035747 kernel: e820: reserve RAM buffer [mem 0x00099800-0x0009ffff] Nov 8 01:37:37.035753 kernel: e820: reserve RAM buffer [mem 0x825be000-0x83ffffff] Nov 8 01:37:37.035758 kernel: e820: reserve RAM buffer [mem 0x8afce000-0x8bffffff] Nov 8 01:37:37.035764 kernel: e820: reserve RAM buffer [mem 0x8c23c000-0x8fffffff] Nov 8 01:37:37.035770 kernel: e820: reserve RAM buffer [mem 0x8ef00000-0x8fffffff] Nov 8 01:37:37.035775 kernel: e820: reserve RAM buffer [mem 0x86f000000-0x86fffffff] Nov 8 01:37:37.035826 kernel: pci 0000:07:00.0: vgaarb: setting as boot VGA device Nov 8 01:37:37.035878 kernel: pci 0000:07:00.0: vgaarb: bridge control possible Nov 8 01:37:37.035932 kernel: pci 0000:07:00.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Nov 8 01:37:37.035941 kernel: vgaarb: loaded Nov 8 01:37:37.035947 kernel: clocksource: Switched to clocksource tsc-early Nov 8 01:37:37.035952 kernel: VFS: Disk quotas dquot_6.6.0 Nov 8 01:37:37.035958 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 8 01:37:37.035964 kernel: pnp: PnP ACPI init Nov 8 01:37:37.036014 kernel: system 00:00: [mem 0x40000000-0x403fffff] has been reserved Nov 8 01:37:37.036062 kernel: pnp 00:02: [dma 0 disabled] Nov 8 01:37:37.036115 kernel: pnp 00:03: [dma 0 disabled] Nov 8 01:37:37.036194 kernel: system 00:04: [io 0x0680-0x069f] has been reserved Nov 8 01:37:37.036260 kernel: system 00:04: [io 0x164e-0x164f] has been reserved Nov 8 01:37:37.036307 kernel: system 00:05: [mem 0xfed10000-0xfed17fff] has been reserved Nov 8 01:37:37.036352 kernel: system 00:05: [mem 0xfed18000-0xfed18fff] has been reserved Nov 8 01:37:37.036396 kernel: system 00:05: [mem 0xfed19000-0xfed19fff] has been reserved Nov 8 01:37:37.036444 kernel: system 00:05: [mem 0xe0000000-0xefffffff] has been reserved Nov 8 01:37:37.036488 kernel: system 00:05: [mem 0xfed20000-0xfed3ffff] has been reserved Nov 8 01:37:37.036535 kernel: system 00:05: [mem 0xfed90000-0xfed93fff] could not be reserved Nov 8 01:37:37.036580 kernel: system 00:05: [mem 0xfed45000-0xfed8ffff] has been reserved Nov 8 01:37:37.036625 kernel: system 00:05: [mem 0xfee00000-0xfeefffff] could not be reserved Nov 8 01:37:37.036674 kernel: system 00:06: [io 0x1800-0x18fe] could not be reserved Nov 8 01:37:37.036720 kernel: system 00:06: [mem 0xfd000000-0xfd69ffff] has been reserved Nov 8 01:37:37.036767 kernel: system 00:06: [mem 0xfd6c0000-0xfd6cffff] has been reserved Nov 8 01:37:37.036811 kernel: system 00:06: [mem 0xfd6f0000-0xfdffffff] has been reserved Nov 8 01:37:37.036856 kernel: system 00:06: [mem 0xfe000000-0xfe01ffff] could not be reserved Nov 8 01:37:37.036899 kernel: system 00:06: [mem 0xfe200000-0xfe7fffff] has been reserved Nov 8 01:37:37.036944 kernel: system 00:06: [mem 0xff000000-0xffffffff] has been reserved Nov 8 01:37:37.036991 kernel: system 00:07: [io 0x2000-0x20fe] has been reserved Nov 8 01:37:37.037000 kernel: pnp: PnP ACPI: found 9 devices Nov 8 01:37:37.037008 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Nov 8 01:37:37.037014 kernel: NET: Registered PF_INET protocol family Nov 8 01:37:37.037019 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 8 01:37:37.037025 kernel: tcp_listen_portaddr_hash hash table entries: 16384 (order: 6, 262144 bytes, linear) Nov 8 01:37:37.037031 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 8 01:37:37.037037 kernel: TCP established hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 8 01:37:37.037043 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Nov 8 01:37:37.037050 kernel: TCP: Hash tables configured (established 262144 bind 65536) Nov 8 01:37:37.037055 kernel: UDP hash table entries: 16384 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.037062 kernel: UDP-Lite hash table entries: 16384 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.037068 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 8 01:37:37.037074 kernel: NET: Registered PF_XDP protocol family Nov 8 01:37:37.037124 kernel: pci 0000:00:15.0: BAR 0: assigned [mem 0x95515000-0x95515fff 64bit] Nov 8 01:37:37.037218 kernel: pci 0000:00:15.1: BAR 0: assigned [mem 0x9551b000-0x9551bfff 64bit] Nov 8 01:37:37.037269 kernel: pci 0000:00:1e.0: BAR 0: assigned [mem 0x9551c000-0x9551cfff 64bit] Nov 8 01:37:37.037320 kernel: pci 0000:01:00.0: BAR 7: no space for [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037371 kernel: pci 0000:01:00.0: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037423 kernel: pci 0000:01:00.1: BAR 7: no space for [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037474 kernel: pci 0000:01:00.1: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037523 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Nov 8 01:37:37.037572 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Nov 8 01:37:37.037622 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Nov 8 01:37:37.037670 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Nov 8 01:37:37.037721 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Nov 8 01:37:37.037771 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Nov 8 01:37:37.037821 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Nov 8 01:37:37.037869 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Nov 8 01:37:37.037918 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Nov 8 01:37:37.037966 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Nov 8 01:37:37.038015 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Nov 8 01:37:37.038067 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Nov 8 01:37:37.038118 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.038210 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.038259 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Nov 8 01:37:37.038308 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.038356 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.038402 kernel: pci_bus 0000:00: Some PCI device resources are unassigned, try booting with pci=realloc Nov 8 01:37:37.038445 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Nov 8 01:37:37.038491 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Nov 8 01:37:37.038534 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Nov 8 01:37:37.038577 kernel: pci_bus 0000:00: resource 7 [mem 0x90000000-0xdfffffff window] Nov 8 01:37:37.038619 kernel: pci_bus 0000:00: resource 8 [mem 0xfc800000-0xfe7fffff window] Nov 8 01:37:37.038670 kernel: pci_bus 0000:01: resource 1 [mem 0x95100000-0x952fffff] Nov 8 01:37:37.038715 kernel: pci_bus 0000:01: resource 2 [mem 0x90000000-0x93ffffff 64bit pref] Nov 8 01:37:37.038765 kernel: pci_bus 0000:03: resource 0 [io 0x5000-0x5fff] Nov 8 01:37:37.038812 kernel: pci_bus 0000:03: resource 1 [mem 0x95400000-0x954fffff] Nov 8 01:37:37.038865 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Nov 8 01:37:37.038910 kernel: pci_bus 0000:04: resource 1 [mem 0x95300000-0x953fffff] Nov 8 01:37:37.038959 kernel: pci_bus 0000:06: resource 0 [io 0x3000-0x3fff] Nov 8 01:37:37.039004 kernel: pci_bus 0000:06: resource 1 [mem 0x94000000-0x950fffff] Nov 8 01:37:37.039051 kernel: pci_bus 0000:07: resource 0 [io 0x3000-0x3fff] Nov 8 01:37:37.039097 kernel: pci_bus 0000:07: resource 1 [mem 0x94000000-0x950fffff] Nov 8 01:37:37.039106 kernel: PCI: CLS 64 bytes, default 64 Nov 8 01:37:37.039113 kernel: DMAR: No ATSR found Nov 8 01:37:37.039118 kernel: DMAR: No SATC found Nov 8 01:37:37.039124 kernel: DMAR: dmar0: Using Queued invalidation Nov 8 01:37:37.039219 kernel: pci 0000:00:00.0: Adding to iommu group 0 Nov 8 01:37:37.039269 kernel: pci 0000:00:01.0: Adding to iommu group 1 Nov 8 01:37:37.039317 kernel: pci 0000:00:08.0: Adding to iommu group 2 Nov 8 01:37:37.039367 kernel: pci 0000:00:12.0: Adding to iommu group 3 Nov 8 01:37:37.039418 kernel: pci 0000:00:14.0: Adding to iommu group 4 Nov 8 01:37:37.039468 kernel: pci 0000:00:14.2: Adding to iommu group 4 Nov 8 01:37:37.039515 kernel: pci 0000:00:15.0: Adding to iommu group 5 Nov 8 01:37:37.039564 kernel: pci 0000:00:15.1: Adding to iommu group 5 Nov 8 01:37:37.039612 kernel: pci 0000:00:16.0: Adding to iommu group 6 Nov 8 01:37:37.039662 kernel: pci 0000:00:16.1: Adding to iommu group 6 Nov 8 01:37:37.039710 kernel: pci 0000:00:16.4: Adding to iommu group 6 Nov 8 01:37:37.039760 kernel: pci 0000:00:17.0: Adding to iommu group 7 Nov 8 01:37:37.039811 kernel: pci 0000:00:1b.0: Adding to iommu group 8 Nov 8 01:37:37.039861 kernel: pci 0000:00:1b.4: Adding to iommu group 9 Nov 8 01:37:37.039909 kernel: pci 0000:00:1b.5: Adding to iommu group 10 Nov 8 01:37:37.039958 kernel: pci 0000:00:1c.0: Adding to iommu group 11 Nov 8 01:37:37.040007 kernel: pci 0000:00:1c.3: Adding to iommu group 12 Nov 8 01:37:37.040055 kernel: pci 0000:00:1e.0: Adding to iommu group 13 Nov 8 01:37:37.040103 kernel: pci 0000:00:1f.0: Adding to iommu group 14 Nov 8 01:37:37.040177 kernel: pci 0000:00:1f.4: Adding to iommu group 14 Nov 8 01:37:37.040246 kernel: pci 0000:00:1f.5: Adding to iommu group 14 Nov 8 01:37:37.040299 kernel: pci 0000:01:00.0: Adding to iommu group 1 Nov 8 01:37:37.040350 kernel: pci 0000:01:00.1: Adding to iommu group 1 Nov 8 01:37:37.040400 kernel: pci 0000:03:00.0: Adding to iommu group 15 Nov 8 01:37:37.040451 kernel: pci 0000:04:00.0: Adding to iommu group 16 Nov 8 01:37:37.040500 kernel: pci 0000:06:00.0: Adding to iommu group 17 Nov 8 01:37:37.040554 kernel: pci 0000:07:00.0: Adding to iommu group 17 Nov 8 01:37:37.040562 kernel: DMAR: Intel(R) Virtualization Technology for Directed I/O Nov 8 01:37:37.040568 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Nov 8 01:37:37.040576 kernel: software IO TLB: mapped [mem 0x0000000086fce000-0x000000008afce000] (64MB) Nov 8 01:37:37.040582 kernel: RAPL PMU: API unit is 2^-32 Joules, 3 fixed counters, 655360 ms ovfl timer Nov 8 01:37:37.040588 kernel: RAPL PMU: hw unit of domain pp0-core 2^-14 Joules Nov 8 01:37:37.040594 kernel: RAPL PMU: hw unit of domain package 2^-14 Joules Nov 8 01:37:37.040599 kernel: RAPL PMU: hw unit of domain dram 2^-14 Joules Nov 8 01:37:37.040651 kernel: platform rtc_cmos: registered platform RTC device (no PNP device found) Nov 8 01:37:37.040660 kernel: Initialise system trusted keyrings Nov 8 01:37:37.040666 kernel: workingset: timestamp_bits=39 max_order=23 bucket_order=0 Nov 8 01:37:37.040673 kernel: Key type asymmetric registered Nov 8 01:37:37.040679 kernel: Asymmetric key parser 'x509' registered Nov 8 01:37:37.040685 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Nov 8 01:37:37.040690 kernel: io scheduler mq-deadline registered Nov 8 01:37:37.040696 kernel: io scheduler kyber registered Nov 8 01:37:37.040702 kernel: io scheduler bfq registered Nov 8 01:37:37.040750 kernel: pcieport 0000:00:01.0: PME: Signaling with IRQ 121 Nov 8 01:37:37.040801 kernel: pcieport 0000:00:1b.0: PME: Signaling with IRQ 122 Nov 8 01:37:37.040850 kernel: pcieport 0000:00:1b.4: PME: Signaling with IRQ 123 Nov 8 01:37:37.040901 kernel: pcieport 0000:00:1b.5: PME: Signaling with IRQ 124 Nov 8 01:37:37.040949 kernel: pcieport 0000:00:1c.0: PME: Signaling with IRQ 125 Nov 8 01:37:37.040998 kernel: pcieport 0000:00:1c.3: PME: Signaling with IRQ 126 Nov 8 01:37:37.041052 kernel: thermal LNXTHERM:00: registered as thermal_zone0 Nov 8 01:37:37.041060 kernel: ACPI: thermal: Thermal Zone [TZ00] (28 C) Nov 8 01:37:37.041066 kernel: ERST: Error Record Serialization Table (ERST) support is initialized. Nov 8 01:37:37.041072 kernel: pstore: Using crash dump compression: deflate Nov 8 01:37:37.041080 kernel: pstore: Registered erst as persistent store backend Nov 8 01:37:37.041085 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Nov 8 01:37:37.041091 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 8 01:37:37.041097 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Nov 8 01:37:37.041103 kernel: 00:03: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Nov 8 01:37:37.041109 kernel: hpet_acpi_add: no address or irqs in _CRS Nov 8 01:37:37.041188 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x1B, rev-id 16) Nov 8 01:37:37.041197 kernel: i8042: PNP: No PS/2 controller found. Nov 8 01:37:37.041262 kernel: rtc_cmos rtc_cmos: RTC can wake from S4 Nov 8 01:37:37.041308 kernel: rtc_cmos rtc_cmos: registered as rtc0 Nov 8 01:37:37.041353 kernel: rtc_cmos rtc_cmos: setting system clock to 2025-11-08T01:37:35 UTC (1762565855) Nov 8 01:37:37.041398 kernel: rtc_cmos rtc_cmos: alarms up to one month, y3k, 114 bytes nvram Nov 8 01:37:37.041406 kernel: intel_pstate: Intel P-state driver initializing Nov 8 01:37:37.041412 kernel: intel_pstate: Disabling energy efficiency optimization Nov 8 01:37:37.041418 kernel: intel_pstate: HWP enabled Nov 8 01:37:37.041424 kernel: vesafb: mode is 1024x768x8, linelength=1024, pages=0 Nov 8 01:37:37.041430 kernel: vesafb: scrolling: redraw Nov 8 01:37:37.041437 kernel: vesafb: Pseudocolor: size=0:8:8:8, shift=0:0:0:0 Nov 8 01:37:37.041443 kernel: vesafb: framebuffer at 0x94000000, mapped to 0x000000009517b72d, using 768k, total 768k Nov 8 01:37:37.041449 kernel: Console: switching to colour frame buffer device 128x48 Nov 8 01:37:37.041454 kernel: fb0: VESA VGA frame buffer device Nov 8 01:37:37.041460 kernel: NET: Registered PF_INET6 protocol family Nov 8 01:37:37.041466 kernel: Segment Routing with IPv6 Nov 8 01:37:37.041472 kernel: In-situ OAM (IOAM) with IPv6 Nov 8 01:37:37.041478 kernel: NET: Registered PF_PACKET protocol family Nov 8 01:37:37.041483 kernel: Key type dns_resolver registered Nov 8 01:37:37.041490 kernel: microcode: Current revision: 0x00000102 Nov 8 01:37:37.041496 kernel: microcode: Microcode Update Driver: v2.2. Nov 8 01:37:37.041501 kernel: IPI shorthand broadcast: enabled Nov 8 01:37:37.041507 kernel: sched_clock: Marking stable (1561085242, 1369070988)->(4401410064, -1471253834) Nov 8 01:37:37.041513 kernel: registered taskstats version 1 Nov 8 01:37:37.041519 kernel: Loading compiled-in X.509 certificates Nov 8 01:37:37.041525 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.113-flatcar: cf7a35a152685ec84a621291e4ce58c959319dfd' Nov 8 01:37:37.041530 kernel: Key type .fscrypt registered Nov 8 01:37:37.041536 kernel: Key type fscrypt-provisioning registered Nov 8 01:37:37.041542 kernel: ima: Allocated hash algorithm: sha1 Nov 8 01:37:37.041548 kernel: ima: No architecture policies found Nov 8 01:37:37.041554 kernel: clk: Disabling unused clocks Nov 8 01:37:37.041560 kernel: Freeing unused kernel image (initmem) memory: 42880K Nov 8 01:37:37.041565 kernel: Write protecting the kernel read-only data: 36864k Nov 8 01:37:37.041571 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Nov 8 01:37:37.041577 kernel: Run /init as init process Nov 8 01:37:37.041583 kernel: with arguments: Nov 8 01:37:37.041589 kernel: /init Nov 8 01:37:37.041595 kernel: with environment: Nov 8 01:37:37.041601 kernel: HOME=/ Nov 8 01:37:37.041606 kernel: TERM=linux Nov 8 01:37:37.041613 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Nov 8 01:37:37.041621 systemd[1]: Detected architecture x86-64. Nov 8 01:37:37.041627 systemd[1]: Running in initrd. Nov 8 01:37:37.041632 systemd[1]: No hostname configured, using default hostname. Nov 8 01:37:37.041639 systemd[1]: Hostname set to . Nov 8 01:37:37.041645 systemd[1]: Initializing machine ID from random generator. Nov 8 01:37:37.041651 systemd[1]: Queued start job for default target initrd.target. Nov 8 01:37:37.041657 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 8 01:37:37.041663 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 8 01:37:37.041669 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Nov 8 01:37:37.041675 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 8 01:37:37.041681 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Nov 8 01:37:37.041688 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Nov 8 01:37:37.041695 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Nov 8 01:37:37.041701 kernel: tsc: Refined TSC clocksource calibration: 3407.998 MHz Nov 8 01:37:37.041707 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Nov 8 01:37:37.041713 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd208cfc, max_idle_ns: 440795283699 ns Nov 8 01:37:37.041719 kernel: clocksource: Switched to clocksource tsc Nov 8 01:37:37.041725 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 8 01:37:37.041732 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 8 01:37:37.041738 systemd[1]: Reached target paths.target - Path Units. Nov 8 01:37:37.041744 systemd[1]: Reached target slices.target - Slice Units. Nov 8 01:37:37.041750 systemd[1]: Reached target swap.target - Swaps. Nov 8 01:37:37.041756 systemd[1]: Reached target timers.target - Timer Units. Nov 8 01:37:37.041762 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Nov 8 01:37:37.041768 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 8 01:37:37.041774 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 8 01:37:37.041780 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Nov 8 01:37:37.041787 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 8 01:37:37.041793 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 8 01:37:37.041799 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 8 01:37:37.041805 systemd[1]: Reached target sockets.target - Socket Units. Nov 8 01:37:37.041811 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Nov 8 01:37:37.041817 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 8 01:37:37.041823 systemd[1]: Finished network-cleanup.service - Network Cleanup. Nov 8 01:37:37.041829 systemd[1]: Starting systemd-fsck-usr.service... Nov 8 01:37:37.041836 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 8 01:37:37.041852 systemd-journald[266]: Collecting audit messages is disabled. Nov 8 01:37:37.041866 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 8 01:37:37.041873 systemd-journald[266]: Journal started Nov 8 01:37:37.041887 systemd-journald[266]: Runtime Journal (/run/log/journal/58bc4733a404490d8afa8ebd88ef6673) is 8.0M, max 639.9M, 631.9M free. Nov 8 01:37:37.055933 systemd-modules-load[267]: Inserted module 'overlay' Nov 8 01:37:37.084246 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 8 01:37:37.132190 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 8 01:37:37.132221 systemd[1]: Started systemd-journald.service - Journal Service. Nov 8 01:37:37.150936 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Nov 8 01:37:37.151028 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 8 01:37:37.151118 systemd[1]: Finished systemd-fsck-usr.service. Nov 8 01:37:37.152231 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 8 01:37:37.169514 systemd-modules-load[267]: Inserted module 'br_netfilter' Nov 8 01:37:37.169924 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 8 01:37:37.258514 kernel: Bridge firewalling registered Nov 8 01:37:37.237732 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 8 01:37:37.259034 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:37:37.290695 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 8 01:37:37.308560 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 8 01:37:37.347466 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 8 01:37:37.348382 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 8 01:37:37.349242 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 8 01:37:37.357312 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 8 01:37:37.357665 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 8 01:37:37.358645 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 8 01:37:37.361373 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:37:37.361943 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Nov 8 01:37:37.376308 systemd-resolved[304]: Positive Trust Anchors: Nov 8 01:37:37.376313 systemd-resolved[304]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 8 01:37:37.376339 systemd-resolved[304]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 8 01:37:37.378002 systemd-resolved[304]: Defaulting to hostname 'linux'. Nov 8 01:37:37.394452 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 8 01:37:37.419443 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 8 01:37:37.524263 dracut-cmdline[306]: dracut-dracut-053 Nov 8 01:37:37.524263 dracut-cmdline[306]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=480a02cf7f2001774aa495c3e719d4173e968e6839485a7d2b207ef2facca472 Nov 8 01:37:37.586201 kernel: SCSI subsystem initialized Nov 8 01:37:37.609178 kernel: Loading iSCSI transport class v2.0-870. Nov 8 01:37:37.632185 kernel: iscsi: registered transport (tcp) Nov 8 01:37:37.665381 kernel: iscsi: registered transport (qla4xxx) Nov 8 01:37:37.665398 kernel: QLogic iSCSI HBA Driver Nov 8 01:37:37.698040 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Nov 8 01:37:37.725410 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Nov 8 01:37:37.781461 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 8 01:37:37.781483 kernel: device-mapper: uevent: version 1.0.3 Nov 8 01:37:37.801219 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Nov 8 01:37:37.859193 kernel: raid6: avx2x4 gen() 53148 MB/s Nov 8 01:37:37.891208 kernel: raid6: avx2x2 gen() 53274 MB/s Nov 8 01:37:37.927589 kernel: raid6: avx2x1 gen() 45177 MB/s Nov 8 01:37:37.927605 kernel: raid6: using algorithm avx2x2 gen() 53274 MB/s Nov 8 01:37:37.975671 kernel: raid6: .... xor() 31128 MB/s, rmw enabled Nov 8 01:37:37.975688 kernel: raid6: using avx2x2 recovery algorithm Nov 8 01:37:38.017183 kernel: xor: automatically using best checksumming function avx Nov 8 01:37:38.135150 kernel: Btrfs loaded, zoned=no, fsverity=no Nov 8 01:37:38.140484 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Nov 8 01:37:38.173538 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 8 01:37:38.180554 systemd-udevd[492]: Using default interface naming scheme 'v255'. Nov 8 01:37:38.184228 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 8 01:37:38.228408 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Nov 8 01:37:38.265848 dracut-pre-trigger[504]: rd.md=0: removing MD RAID activation Nov 8 01:37:38.282972 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Nov 8 01:37:38.313520 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 8 01:37:38.414997 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 8 01:37:38.449206 kernel: pps_core: LinuxPPS API ver. 1 registered Nov 8 01:37:38.449264 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Nov 8 01:37:38.477184 kernel: cryptd: max_cpu_qlen set to 1000 Nov 8 01:37:38.500821 kernel: ACPI: bus type USB registered Nov 8 01:37:38.500840 kernel: usbcore: registered new interface driver usbfs Nov 8 01:37:38.516025 kernel: usbcore: registered new interface driver hub Nov 8 01:37:38.530815 kernel: usbcore: registered new device driver usb Nov 8 01:37:38.546145 kernel: PTP clock support registered Nov 8 01:37:38.546163 kernel: libata version 3.00 loaded. Nov 8 01:37:38.562812 kernel: ahci 0000:00:17.0: version 3.0 Nov 8 01:37:38.562985 kernel: AVX2 version of gcm_enc/dec engaged. Nov 8 01:37:38.564466 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Nov 8 01:37:38.641566 kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 7 ports 6 Gbps 0x7f impl SATA mode Nov 8 01:37:38.641786 kernel: ahci 0000:00:17.0: flags: 64bit ncq sntf clo only pio slum part ems deso sadm sds apst Nov 8 01:37:38.641954 kernel: scsi host0: ahci Nov 8 01:37:38.642132 kernel: scsi host1: ahci Nov 8 01:37:38.642326 kernel: scsi host2: ahci Nov 8 01:37:38.627515 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Nov 8 01:37:38.669246 kernel: scsi host3: ahci Nov 8 01:37:38.669526 kernel: scsi host4: ahci Nov 8 01:37:38.656853 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Nov 8 01:37:38.715796 kernel: scsi host5: ahci Nov 8 01:37:38.716140 kernel: scsi host6: ahci Nov 8 01:37:38.716451 kernel: ata1: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516100 irq 127 Nov 8 01:37:38.685827 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 8 01:37:38.850246 kernel: ata2: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516180 irq 127 Nov 8 01:37:38.850260 kernel: ata3: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516200 irq 127 Nov 8 01:37:38.850268 kernel: ata4: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516280 irq 127 Nov 8 01:37:38.850276 kernel: ata5: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516300 irq 127 Nov 8 01:37:38.850283 kernel: ata6: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516380 irq 127 Nov 8 01:37:38.850290 kernel: ata7: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516400 irq 127 Nov 8 01:37:38.850298 kernel: AES CTR mode by8 optimization enabled Nov 8 01:37:38.850305 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Nov 8 01:37:38.836242 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 8 01:37:38.875312 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Nov 8 01:37:38.875247 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 8 01:37:38.953190 kernel: igb 0000:03:00.0: added PHC on eth0 Nov 8 01:37:38.953280 kernel: igb 0000:03:00.0: Intel(R) Gigabit Ethernet Network Connection Nov 8 01:37:38.953357 kernel: igb 0000:03:00.0: eth0: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:70:d3:7e Nov 8 01:37:38.953430 kernel: igb 0000:03:00.0: eth0: PBA No: 010000-000 Nov 8 01:37:38.953500 kernel: igb 0000:03:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Nov 8 01:37:38.875280 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:37:39.068988 kernel: mlx5_core 0000:01:00.0: firmware version: 14.31.1014 Nov 8 01:37:39.069080 kernel: mlx5_core 0000:01:00.0: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Nov 8 01:37:39.069159 kernel: igb 0000:04:00.0: added PHC on eth1 Nov 8 01:37:39.069231 kernel: igb 0000:04:00.0: Intel(R) Gigabit Ethernet Network Connection Nov 8 01:37:39.069297 kernel: igb 0000:04:00.0: eth1: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:70:d3:7f Nov 8 01:37:39.069362 kernel: igb 0000:04:00.0: eth1: PBA No: 010000-000 Nov 8 01:37:39.069427 kernel: ata3: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.069436 kernel: igb 0000:04:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Nov 8 01:37:39.069499 kernel: ata5: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:38.969257 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 8 01:37:39.096599 kernel: ata2: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Nov 8 01:37:39.096612 kernel: ata4: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.111144 kernel: ata1: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Nov 8 01:37:39.126145 kernel: ata6: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.139177 kernel: ata7: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.152197 kernel: ata2.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Nov 8 01:37:39.166181 kernel: ata1.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Nov 8 01:37:39.182274 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Nov 8 01:37:39.235500 kernel: ata2.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Nov 8 01:37:39.235512 kernel: ata1.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Nov 8 01:37:39.235520 kernel: ata1.00: Features: NCQ-prio Nov 8 01:37:39.235528 kernel: ata2.00: Features: NCQ-prio Nov 8 01:37:39.198239 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 8 01:37:39.347213 kernel: ata1.00: configured for UDMA/133 Nov 8 01:37:39.347232 kernel: ata2.00: configured for UDMA/133 Nov 8 01:37:39.347242 kernel: scsi 0:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Nov 8 01:37:39.347348 kernel: mlx5_core 0000:01:00.0: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Nov 8 01:37:39.347438 kernel: mlx5_core 0000:01:00.0: Port module event: module 0, Cable plugged Nov 8 01:37:39.347516 kernel: scsi 1:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Nov 8 01:37:39.347595 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Nov 8 01:37:39.347677 kernel: igb 0000:04:00.0 eno2: renamed from eth1 Nov 8 01:37:39.347755 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 1 Nov 8 01:37:39.198269 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:37:39.252741 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Nov 8 01:37:39.376146 kernel: xhci_hcd 0000:00:14.0: hcc params 0x200077c1 hci version 0x110 quirks 0x0000000000009810 Nov 8 01:37:39.376286 kernel: igb 0000:03:00.0 eno1: renamed from eth0 Nov 8 01:37:39.381201 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Nov 8 01:37:39.418234 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 2 Nov 8 01:37:39.427183 kernel: xhci_hcd 0000:00:14.0: Host supports USB 3.1 Enhanced SuperSpeed Nov 8 01:37:39.427284 kernel: hub 1-0:1.0: USB hub found Nov 8 01:37:39.459060 kernel: hub 1-0:1.0: 16 ports detected Nov 8 01:37:39.472226 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 8 01:37:39.513641 kernel: hub 2-0:1.0: USB hub found Nov 8 01:37:39.513788 kernel: hub 2-0:1.0: 10 ports detected Nov 8 01:37:39.513867 kernel: mlx5_core 0000:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Nov 8 01:37:39.494424 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Nov 8 01:37:39.764515 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:39.764529 kernel: mlx5_core 0000:01:00.1: firmware version: 14.31.1014 Nov 8 01:37:39.764619 kernel: ata2.00: Enabling discard_zeroes_data Nov 8 01:37:39.764628 kernel: sd 0:0:0:0: [sda] 937703088 512-byte logical blocks: (480 GB/447 GiB) Nov 8 01:37:39.764703 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Nov 8 01:37:39.764767 kernel: mlx5_core 0000:01:00.1: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Nov 8 01:37:39.764835 kernel: sd 0:0:0:0: [sda] Write Protect is off Nov 8 01:37:39.764902 kernel: sd 1:0:0:0: [sdb] 937703088 512-byte logical blocks: (480 GB/447 GiB) Nov 8 01:37:39.764969 kernel: sd 1:0:0:0: [sdb] 4096-byte physical blocks Nov 8 01:37:39.765031 kernel: sd 1:0:0:0: [sdb] Write Protect is off Nov 8 01:37:39.765093 kernel: sd 1:0:0:0: [sdb] Mode Sense: 00 3a 00 00 Nov 8 01:37:39.765164 kernel: sd 1:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Nov 8 01:37:39.765257 kernel: sd 1:0:0:0: [sdb] Preferred minimum I/O size 4096 bytes Nov 8 01:37:39.765331 kernel: ata2.00: Enabling discard_zeroes_data Nov 8 01:37:39.765343 kernel: sd 1:0:0:0: [sdb] Attached SCSI disk Nov 8 01:37:39.765410 kernel: usb 1-14: new high-speed USB device number 2 using xhci_hcd Nov 8 01:37:39.765427 kernel: sd 0:0:0:0: [sda] Mode Sense: 00 3a 00 00 Nov 8 01:37:39.765489 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Nov 8 01:37:39.765552 kernel: sd 0:0:0:0: [sda] Preferred minimum I/O size 4096 bytes Nov 8 01:37:39.796951 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:39.821594 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 8 01:37:39.821611 kernel: GPT:9289727 != 937703087 Nov 8 01:37:39.837602 kernel: GPT:Alternate GPT header not at the end of the disk. Nov 8 01:37:39.837618 kernel: mlx5_core 0000:01:00.1: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Nov 8 01:37:39.837702 kernel: GPT:9289727 != 937703087 Nov 8 01:37:39.837711 kernel: GPT: Use GNU Parted to correct GPT errors. Nov 8 01:37:39.837718 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:39.855147 kernel: mlx5_core 0000:01:00.1: Port module event: module 1, Cable plugged Nov 8 01:37:39.855238 kernel: hub 1-14:1.0: USB hub found Nov 8 01:37:39.855323 kernel: hub 1-14:1.0: 4 ports detected Nov 8 01:37:39.864794 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Nov 8 01:37:39.970174 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:37:40.006853 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Micron_5300_MTFDDAK480TDT EFI-SYSTEM. Nov 8 01:37:40.071401 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by (udev-worker) (669) Nov 8 01:37:40.071417 kernel: BTRFS: device fsid a2737782-a37e-42f9-8b56-489a87f47acc devid 1 transid 35 /dev/sda3 scanned by (udev-worker) (663) Nov 8 01:37:40.061950 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Micron_5300_MTFDDAK480TDT ROOT. Nov 8 01:37:40.085985 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Nov 8 01:37:40.118211 kernel: mlx5_core 0000:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Nov 8 01:37:40.134191 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Micron_5300_MTFDDAK480TDT USR-A. Nov 8 01:37:40.203413 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: renamed from eth1 Nov 8 01:37:40.203508 kernel: usb 1-14.1: new low-speed USB device number 3 using xhci_hcd Nov 8 01:37:40.203526 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: renamed from eth0 Nov 8 01:37:40.156264 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Micron_5300_MTFDDAK480TDT USR-A. Nov 8 01:37:40.205451 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Nov 8 01:37:40.261194 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:40.261208 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:40.261219 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:40.237672 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 8 01:37:40.262216 disk-uuid[723]: Primary Header is updated. Nov 8 01:37:40.262216 disk-uuid[723]: Secondary Entries is updated. Nov 8 01:37:40.262216 disk-uuid[723]: Secondary Header is updated. Nov 8 01:37:40.315596 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:40.315607 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:40.315614 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:40.340149 kernel: hid: raw HID events driver (C) Jiri Kosina Nov 8 01:37:40.363274 kernel: usbcore: registered new interface driver usbhid Nov 8 01:37:40.363348 kernel: usbhid: USB HID core driver Nov 8 01:37:40.410286 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.0/0003:0557:2419.0001/input/input0 Nov 8 01:37:40.423624 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:37:40.508665 kernel: hid-generic 0003:0557:2419.0001: input,hidraw0: USB HID v1.00 Keyboard [HID 0557:2419] on usb-0000:00:14.0-14.1/input0 Nov 8 01:37:40.508790 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.1/0003:0557:2419.0002/input/input1 Nov 8 01:37:40.544027 kernel: hid-generic 0003:0557:2419.0002: input,hidraw1: USB HID v1.00 Mouse [HID 0557:2419] on usb-0000:00:14.0-14.1/input1 Nov 8 01:37:41.297131 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:41.317129 disk-uuid[724]: The operation has completed successfully. Nov 8 01:37:41.325257 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:41.354135 systemd[1]: disk-uuid.service: Deactivated successfully. Nov 8 01:37:41.354192 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Nov 8 01:37:41.393386 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Nov 8 01:37:41.418325 sh[751]: Success Nov 8 01:37:41.428245 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Nov 8 01:37:41.474509 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Nov 8 01:37:41.495293 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Nov 8 01:37:41.503544 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Nov 8 01:37:41.565055 kernel: BTRFS info (device dm-0): first mount of filesystem a2737782-a37e-42f9-8b56-489a87f47acc Nov 8 01:37:41.565075 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:41.587034 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Nov 8 01:37:41.606571 kernel: BTRFS info (device dm-0): disabling log replay at mount time Nov 8 01:37:41.624983 kernel: BTRFS info (device dm-0): using free space tree Nov 8 01:37:41.663190 kernel: BTRFS info (device dm-0): enabling ssd optimizations Nov 8 01:37:41.665003 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Nov 8 01:37:41.673594 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Nov 8 01:37:41.688314 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Nov 8 01:37:41.700967 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Nov 8 01:37:41.812605 kernel: BTRFS info (device sda6): first mount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:41.812631 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:41.812646 kernel: BTRFS info (device sda6): using free space tree Nov 8 01:37:41.812654 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 8 01:37:41.812661 kernel: BTRFS info (device sda6): auto enabling async discard Nov 8 01:37:41.836223 kernel: BTRFS info (device sda6): last unmount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:41.845330 systemd[1]: Finished ignition-setup.service - Ignition (setup). Nov 8 01:37:41.874329 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Nov 8 01:37:41.891801 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 8 01:37:41.920271 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 8 01:37:41.931824 systemd-networkd[934]: lo: Link UP Nov 8 01:37:41.931827 systemd-networkd[934]: lo: Gained carrier Nov 8 01:37:41.940648 ignition[859]: Ignition 2.19.0 Nov 8 01:37:41.935073 systemd-networkd[934]: Enumeration completed Nov 8 01:37:41.940653 ignition[859]: Stage: fetch-offline Nov 8 01:37:41.935149 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 8 01:37:41.940675 ignition[859]: no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:41.936000 systemd-networkd[934]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:41.940680 ignition[859]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:41.942929 unknown[859]: fetched base config from "system" Nov 8 01:37:41.940733 ignition[859]: parsed url from cmdline: "" Nov 8 01:37:41.942933 unknown[859]: fetched user config from "system" Nov 8 01:37:41.940735 ignition[859]: no config URL provided Nov 8 01:37:41.947571 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Nov 8 01:37:41.940737 ignition[859]: reading system config file "/usr/lib/ignition/user.ign" Nov 8 01:37:41.964887 systemd-networkd[934]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:41.940761 ignition[859]: parsing config with SHA512: 73ad4a693266b007874f0d28e1091c13061053cdd7b48746fdbadf195e470ff304d4cbb2e51232992df41ca9e71db2bb394fceb2ae45c785dd19a5ae480522f7 Nov 8 01:37:41.966649 systemd[1]: Reached target network.target - Network. Nov 8 01:37:41.943204 ignition[859]: fetch-offline: fetch-offline passed Nov 8 01:37:41.981323 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Nov 8 01:37:41.943206 ignition[859]: POST message to Packet Timeline Nov 8 01:37:41.990324 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Nov 8 01:37:41.943209 ignition[859]: POST Status error: resource requires networking Nov 8 01:37:41.993436 systemd-networkd[934]: enp1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:41.943259 ignition[859]: Ignition finished successfully Nov 8 01:37:42.001986 ignition[947]: Ignition 2.19.0 Nov 8 01:37:42.001994 ignition[947]: Stage: kargs Nov 8 01:37:42.002201 ignition[947]: no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:42.002215 ignition[947]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:42.003345 ignition[947]: kargs: kargs passed Nov 8 01:37:42.227291 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Nov 8 01:37:42.221665 systemd-networkd[934]: enp1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:42.003350 ignition[947]: POST message to Packet Timeline Nov 8 01:37:42.003366 ignition[947]: GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:37:42.004237 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:41411->[::1]:53: read: connection refused Nov 8 01:37:42.204476 ignition[947]: GET https://metadata.packet.net/metadata: attempt #2 Nov 8 01:37:42.204969 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:55986->[::1]:53: read: connection refused Nov 8 01:37:42.510193 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Nov 8 01:37:42.510825 systemd-networkd[934]: eno1: Link UP Nov 8 01:37:42.511046 systemd-networkd[934]: eno2: Link UP Nov 8 01:37:42.511222 systemd-networkd[934]: enp1s0f0np0: Link UP Nov 8 01:37:42.511413 systemd-networkd[934]: enp1s0f0np0: Gained carrier Nov 8 01:37:42.520425 systemd-networkd[934]: enp1s0f1np1: Link UP Nov 8 01:37:42.549313 systemd-networkd[934]: enp1s0f0np0: DHCPv4 address 139.178.94.189/31, gateway 139.178.94.188 acquired from 145.40.83.140 Nov 8 01:37:42.605240 ignition[947]: GET https://metadata.packet.net/metadata: attempt #3 Nov 8 01:37:42.606374 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:52706->[::1]:53: read: connection refused Nov 8 01:37:43.265883 systemd-networkd[934]: enp1s0f1np1: Gained carrier Nov 8 01:37:43.407567 ignition[947]: GET https://metadata.packet.net/metadata: attempt #4 Nov 8 01:37:43.408720 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:42734->[::1]:53: read: connection refused Nov 8 01:37:43.649767 systemd-networkd[934]: enp1s0f0np0: Gained IPv6LL Nov 8 01:37:45.010493 ignition[947]: GET https://metadata.packet.net/metadata: attempt #5 Nov 8 01:37:45.011603 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:46313->[::1]:53: read: connection refused Nov 8 01:37:45.121758 systemd-networkd[934]: enp1s0f1np1: Gained IPv6LL Nov 8 01:37:48.215234 ignition[947]: GET https://metadata.packet.net/metadata: attempt #6 Nov 8 01:37:52.062805 ignition[947]: GET result: OK Nov 8 01:37:53.303309 ignition[947]: Ignition finished successfully Nov 8 01:37:53.308523 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Nov 8 01:37:53.337391 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Nov 8 01:37:53.343585 ignition[966]: Ignition 2.19.0 Nov 8 01:37:53.343589 ignition[966]: Stage: disks Nov 8 01:37:53.343702 ignition[966]: no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:53.343709 ignition[966]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:53.344284 ignition[966]: disks: disks passed Nov 8 01:37:53.344286 ignition[966]: POST message to Packet Timeline Nov 8 01:37:53.344296 ignition[966]: GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:37:54.760482 ignition[966]: GET result: OK Nov 8 01:37:55.287621 ignition[966]: Ignition finished successfully Nov 8 01:37:55.290803 systemd[1]: Finished ignition-disks.service - Ignition (disks). Nov 8 01:37:55.306407 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Nov 8 01:37:55.324410 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 8 01:37:55.345405 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 8 01:37:55.366564 systemd[1]: Reached target sysinit.target - System Initialization. Nov 8 01:37:55.386440 systemd[1]: Reached target basic.target - Basic System. Nov 8 01:37:55.415418 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Nov 8 01:37:55.455093 systemd-fsck[984]: ROOT: clean, 14/553520 files, 52654/553472 blocks Nov 8 01:37:55.465598 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Nov 8 01:37:55.501409 systemd[1]: Mounting sysroot.mount - /sysroot... Nov 8 01:37:55.603152 kernel: EXT4-fs (sda9): mounted filesystem 3cd35b5c-4e0e-45c1-abc9-cf70eebd42df r/w with ordered data mode. Quota mode: none. Nov 8 01:37:55.603666 systemd[1]: Mounted sysroot.mount - /sysroot. Nov 8 01:37:55.613571 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Nov 8 01:37:55.639354 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 8 01:37:55.639927 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Nov 8 01:37:55.662921 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Nov 8 01:37:55.685146 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (993) Nov 8 01:37:55.709164 kernel: BTRFS info (device sda6): first mount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:55.709200 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:55.709208 kernel: BTRFS info (device sda6): using free space tree Nov 8 01:37:55.734696 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Nov 8 01:37:55.798447 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 8 01:37:55.798462 kernel: BTRFS info (device sda6): auto enabling async discard Nov 8 01:37:55.798470 coreos-metadata[995]: Nov 08 01:37:55.738 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Nov 8 01:37:55.798378 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Nov 8 01:37:55.798396 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Nov 8 01:37:55.869353 coreos-metadata[1013]: Nov 08 01:37:55.854 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Nov 8 01:37:55.819509 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 8 01:37:55.858412 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Nov 8 01:37:55.894411 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Nov 8 01:37:55.925514 initrd-setup-root[1025]: cut: /sysroot/etc/passwd: No such file or directory Nov 8 01:37:55.935257 initrd-setup-root[1032]: cut: /sysroot/etc/group: No such file or directory Nov 8 01:37:55.945478 initrd-setup-root[1039]: cut: /sysroot/etc/shadow: No such file or directory Nov 8 01:37:55.955206 initrd-setup-root[1046]: cut: /sysroot/etc/gshadow: No such file or directory Nov 8 01:37:55.991664 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Nov 8 01:37:56.015417 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Nov 8 01:37:56.051360 kernel: BTRFS info (device sda6): last unmount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:56.033783 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Nov 8 01:37:56.059918 systemd[1]: sysroot-oem.mount: Deactivated successfully. Nov 8 01:37:56.081380 ignition[1114]: INFO : Ignition 2.19.0 Nov 8 01:37:56.081380 ignition[1114]: INFO : Stage: mount Nov 8 01:37:56.095377 ignition[1114]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:56.095377 ignition[1114]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:56.095377 ignition[1114]: INFO : mount: mount passed Nov 8 01:37:56.095377 ignition[1114]: INFO : POST message to Packet Timeline Nov 8 01:37:56.095377 ignition[1114]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:37:56.089479 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Nov 8 01:37:56.654305 coreos-metadata[995]: Nov 08 01:37:56.654 INFO Fetch successful Nov 8 01:37:56.734943 coreos-metadata[995]: Nov 08 01:37:56.734 INFO wrote hostname ci-4081.3.6-n-3863d1ecf5 to /sysroot/etc/hostname Nov 8 01:37:56.736459 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Nov 8 01:37:56.803652 coreos-metadata[1013]: Nov 08 01:37:56.803 INFO Fetch successful Nov 8 01:37:56.884580 systemd[1]: flatcar-static-network.service: Deactivated successfully. Nov 8 01:37:56.884639 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Nov 8 01:37:57.156290 ignition[1114]: INFO : GET result: OK Nov 8 01:37:58.389560 ignition[1114]: INFO : Ignition finished successfully Nov 8 01:37:58.393418 systemd[1]: Finished ignition-mount.service - Ignition (mount). Nov 8 01:37:58.423359 systemd[1]: Starting ignition-files.service - Ignition (files)... Nov 8 01:37:58.433372 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 8 01:37:58.488200 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (1137) Nov 8 01:37:58.518038 kernel: BTRFS info (device sda6): first mount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:58.518054 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:58.536209 kernel: BTRFS info (device sda6): using free space tree Nov 8 01:37:58.575522 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 8 01:37:58.575538 kernel: BTRFS info (device sda6): auto enabling async discard Nov 8 01:37:58.589523 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 8 01:37:58.618642 ignition[1154]: INFO : Ignition 2.19.0 Nov 8 01:37:58.618642 ignition[1154]: INFO : Stage: files Nov 8 01:37:58.633385 ignition[1154]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:58.633385 ignition[1154]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:58.633385 ignition[1154]: DEBUG : files: compiled without relabeling support, skipping Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Nov 8 01:37:58.633385 ignition[1154]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Nov 8 01:37:58.622968 unknown[1154]: wrote ssh authorized keys file for user: core Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:59.046404 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Nov 8 01:37:59.181257 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Nov 8 01:37:59.574837 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:59.574837 ignition[1154]: INFO : files: op(c): [started] processing unit "containerd.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(c): [finished] processing unit "containerd.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: files passed Nov 8 01:37:59.605414 ignition[1154]: INFO : POST message to Packet Timeline Nov 8 01:37:59.605414 ignition[1154]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:38:00.621738 ignition[1154]: INFO : GET result: OK Nov 8 01:38:01.755074 ignition[1154]: INFO : Ignition finished successfully Nov 8 01:38:01.758235 systemd[1]: Finished ignition-files.service - Ignition (files). Nov 8 01:38:01.797425 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Nov 8 01:38:01.807764 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Nov 8 01:38:01.828696 systemd[1]: ignition-quench.service: Deactivated successfully. Nov 8 01:38:01.828787 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Nov 8 01:38:01.868103 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 8 01:38:01.880654 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Nov 8 01:38:01.912370 initrd-setup-root-after-ignition[1193]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 8 01:38:01.912370 initrd-setup-root-after-ignition[1193]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Nov 8 01:38:01.926360 initrd-setup-root-after-ignition[1197]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 8 01:38:01.919359 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Nov 8 01:38:01.978827 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 8 01:38:01.978883 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Nov 8 01:38:01.997642 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Nov 8 01:38:02.018314 systemd[1]: Reached target initrd.target - Initrd Default Target. Nov 8 01:38:02.042612 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Nov 8 01:38:02.053519 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Nov 8 01:38:02.119772 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 8 01:38:02.145566 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Nov 8 01:38:02.209436 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Nov 8 01:38:02.221772 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 8 01:38:02.242844 systemd[1]: Stopped target timers.target - Timer Units. Nov 8 01:38:02.263876 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 8 01:38:02.264313 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 8 01:38:02.292983 systemd[1]: Stopped target initrd.target - Initrd Default Target. Nov 8 01:38:02.313783 systemd[1]: Stopped target basic.target - Basic System. Nov 8 01:38:02.334869 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Nov 8 01:38:02.355867 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Nov 8 01:38:02.376759 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Nov 8 01:38:02.397777 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Nov 8 01:38:02.417876 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Nov 8 01:38:02.438796 systemd[1]: Stopped target sysinit.target - System Initialization. Nov 8 01:38:02.459888 systemd[1]: Stopped target local-fs.target - Local File Systems. Nov 8 01:38:02.479866 systemd[1]: Stopped target swap.target - Swaps. Nov 8 01:38:02.497748 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 8 01:38:02.498181 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Nov 8 01:38:02.524885 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Nov 8 01:38:02.545887 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 8 01:38:02.566635 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Nov 8 01:38:02.567087 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 8 01:38:02.588648 systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 8 01:38:02.589044 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Nov 8 01:38:02.620747 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Nov 8 01:38:02.621215 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Nov 8 01:38:02.640963 systemd[1]: Stopped target paths.target - Path Units. Nov 8 01:38:02.660620 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 8 01:38:02.661061 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 8 01:38:02.683778 systemd[1]: Stopped target slices.target - Slice Units. Nov 8 01:38:02.701765 systemd[1]: Stopped target sockets.target - Socket Units. Nov 8 01:38:02.721831 systemd[1]: iscsid.socket: Deactivated successfully. Nov 8 01:38:02.722136 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Nov 8 01:38:02.741894 systemd[1]: iscsiuio.socket: Deactivated successfully. Nov 8 01:38:02.742226 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 8 01:38:02.764966 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Nov 8 01:38:02.765399 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 8 01:38:02.784838 systemd[1]: ignition-files.service: Deactivated successfully. Nov 8 01:38:02.892329 ignition[1217]: INFO : Ignition 2.19.0 Nov 8 01:38:02.892329 ignition[1217]: INFO : Stage: umount Nov 8 01:38:02.892329 ignition[1217]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 8 01:38:02.892329 ignition[1217]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:38:02.892329 ignition[1217]: INFO : umount: umount passed Nov 8 01:38:02.892329 ignition[1217]: INFO : POST message to Packet Timeline Nov 8 01:38:02.892329 ignition[1217]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:38:02.785242 systemd[1]: Stopped ignition-files.service - Ignition (files). Nov 8 01:38:02.802847 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Nov 8 01:38:02.803261 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Nov 8 01:38:02.832423 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Nov 8 01:38:02.864407 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Nov 8 01:38:02.883355 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 8 01:38:02.883447 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Nov 8 01:38:02.903418 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 8 01:38:02.903517 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Nov 8 01:38:02.947284 systemd[1]: sysroot-boot.mount: Deactivated successfully. Nov 8 01:38:02.949180 systemd[1]: sysroot-boot.service: Deactivated successfully. Nov 8 01:38:02.949441 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Nov 8 01:38:02.964076 systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 8 01:38:02.964336 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Nov 8 01:38:03.872363 ignition[1217]: INFO : GET result: OK Nov 8 01:38:04.302260 ignition[1217]: INFO : Ignition finished successfully Nov 8 01:38:04.305340 systemd[1]: ignition-mount.service: Deactivated successfully. Nov 8 01:38:04.305647 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Nov 8 01:38:04.323400 systemd[1]: Stopped target network.target - Network. Nov 8 01:38:04.338392 systemd[1]: ignition-disks.service: Deactivated successfully. Nov 8 01:38:04.338568 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Nov 8 01:38:04.357491 systemd[1]: ignition-kargs.service: Deactivated successfully. Nov 8 01:38:04.357628 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Nov 8 01:38:04.377533 systemd[1]: ignition-setup.service: Deactivated successfully. Nov 8 01:38:04.377695 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Nov 8 01:38:04.396664 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Nov 8 01:38:04.396828 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Nov 8 01:38:04.415666 systemd[1]: initrd-setup-root.service: Deactivated successfully. Nov 8 01:38:04.415831 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Nov 8 01:38:04.435047 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Nov 8 01:38:04.451300 systemd-networkd[934]: enp1s0f0np0: DHCPv6 lease lost Nov 8 01:38:04.454625 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Nov 8 01:38:04.462369 systemd-networkd[934]: enp1s0f1np1: DHCPv6 lease lost Nov 8 01:38:04.473122 systemd[1]: systemd-resolved.service: Deactivated successfully. Nov 8 01:38:04.473439 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Nov 8 01:38:04.492671 systemd[1]: systemd-networkd.service: Deactivated successfully. Nov 8 01:38:04.492987 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Nov 8 01:38:04.513414 systemd[1]: systemd-networkd.socket: Deactivated successfully. Nov 8 01:38:04.513542 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Nov 8 01:38:04.545473 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Nov 8 01:38:04.552535 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Nov 8 01:38:04.552709 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 8 01:38:04.579578 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 8 01:38:04.579744 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 8 01:38:04.598572 systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 8 01:38:04.598743 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Nov 8 01:38:04.618557 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 8 01:38:04.618739 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 8 01:38:04.637820 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 8 01:38:04.659434 systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 8 01:38:04.659850 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 8 01:38:04.689707 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 8 01:38:04.689744 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Nov 8 01:38:04.715249 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 8 01:38:04.715275 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Nov 8 01:38:04.735356 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 8 01:38:04.735425 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Nov 8 01:38:04.778353 systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 8 01:38:04.778508 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Nov 8 01:38:04.816342 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 8 01:38:04.816495 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:38:04.870264 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Nov 8 01:38:04.899257 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 8 01:38:04.899411 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 8 01:38:04.920553 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Nov 8 01:38:04.920703 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 8 01:38:05.166378 systemd-journald[266]: Received SIGTERM from PID 1 (systemd). Nov 8 01:38:04.941540 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 8 01:38:04.941687 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Nov 8 01:38:04.962480 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 8 01:38:04.962652 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:38:04.985614 systemd[1]: network-cleanup.service: Deactivated successfully. Nov 8 01:38:04.985920 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Nov 8 01:38:05.003378 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 8 01:38:05.003641 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Nov 8 01:38:05.025625 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Nov 8 01:38:05.061630 systemd[1]: Starting initrd-switch-root.service - Switch Root... Nov 8 01:38:05.095653 systemd[1]: Switching root. Nov 8 01:38:05.270338 systemd-journald[266]: Journal stopped Nov 8 01:37:37.027764 kernel: Linux version 6.6.113-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Nov 7 22:45:04 -00 2025 Nov 8 01:37:37.027778 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=480a02cf7f2001774aa495c3e719d4173e968e6839485a7d2b207ef2facca472 Nov 8 01:37:37.027785 kernel: BIOS-provided physical RAM map: Nov 8 01:37:37.027789 kernel: BIOS-e820: [mem 0x0000000000000000-0x00000000000997ff] usable Nov 8 01:37:37.027793 kernel: BIOS-e820: [mem 0x0000000000099800-0x000000000009ffff] reserved Nov 8 01:37:37.027797 kernel: BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved Nov 8 01:37:37.027802 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003fffffff] usable Nov 8 01:37:37.027806 kernel: BIOS-e820: [mem 0x0000000040000000-0x00000000403fffff] reserved Nov 8 01:37:37.027810 kernel: BIOS-e820: [mem 0x0000000040400000-0x00000000825bdfff] usable Nov 8 01:37:37.027814 kernel: BIOS-e820: [mem 0x00000000825be000-0x00000000825befff] ACPI NVS Nov 8 01:37:37.027819 kernel: BIOS-e820: [mem 0x00000000825bf000-0x00000000825bffff] reserved Nov 8 01:37:37.027824 kernel: BIOS-e820: [mem 0x00000000825c0000-0x000000008afcdfff] usable Nov 8 01:37:37.027828 kernel: BIOS-e820: [mem 0x000000008afce000-0x000000008c0b2fff] reserved Nov 8 01:37:37.027832 kernel: BIOS-e820: [mem 0x000000008c0b3000-0x000000008c23bfff] usable Nov 8 01:37:37.027837 kernel: BIOS-e820: [mem 0x000000008c23c000-0x000000008c66dfff] ACPI NVS Nov 8 01:37:37.027842 kernel: BIOS-e820: [mem 0x000000008c66e000-0x000000008eefefff] reserved Nov 8 01:37:37.027848 kernel: BIOS-e820: [mem 0x000000008eeff000-0x000000008eefffff] usable Nov 8 01:37:37.027852 kernel: BIOS-e820: [mem 0x000000008ef00000-0x000000008fffffff] reserved Nov 8 01:37:37.027857 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Nov 8 01:37:37.027862 kernel: BIOS-e820: [mem 0x00000000fe000000-0x00000000fe010fff] reserved Nov 8 01:37:37.027867 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec00fff] reserved Nov 8 01:37:37.027871 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Nov 8 01:37:37.027876 kernel: BIOS-e820: [mem 0x00000000ff000000-0x00000000ffffffff] reserved Nov 8 01:37:37.027881 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000086effffff] usable Nov 8 01:37:37.027885 kernel: NX (Execute Disable) protection: active Nov 8 01:37:37.027890 kernel: APIC: Static calls initialized Nov 8 01:37:37.027895 kernel: SMBIOS 3.2.1 present. Nov 8 01:37:37.027900 kernel: DMI: Supermicro SYS-5019C-MR-PH004/X11SCM-F, BIOS 2.6 12/03/2024 Nov 8 01:37:37.027906 kernel: tsc: Detected 3400.000 MHz processor Nov 8 01:37:37.027910 kernel: tsc: Detected 3399.906 MHz TSC Nov 8 01:37:37.027915 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Nov 8 01:37:37.027921 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Nov 8 01:37:37.027926 kernel: last_pfn = 0x86f000 max_arch_pfn = 0x400000000 Nov 8 01:37:37.027930 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 23), built from 10 variable MTRRs Nov 8 01:37:37.027935 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Nov 8 01:37:37.027940 kernel: last_pfn = 0x8ef00 max_arch_pfn = 0x400000000 Nov 8 01:37:37.027945 kernel: Using GB pages for direct mapping Nov 8 01:37:37.027951 kernel: ACPI: Early table checksum verification disabled Nov 8 01:37:37.027955 kernel: ACPI: RSDP 0x00000000000F05B0 000024 (v02 SUPERM) Nov 8 01:37:37.027960 kernel: ACPI: XSDT 0x000000008C54F0C8 00010C (v01 SUPERM SUPERM 01072009 AMI 00010013) Nov 8 01:37:37.027967 kernel: ACPI: FACP 0x000000008C58B670 000114 (v06 01072009 AMI 00010013) Nov 8 01:37:37.027972 kernel: ACPI: DSDT 0x000000008C54F268 03C404 (v02 SUPERM SMCI--MB 01072009 INTL 20160527) Nov 8 01:37:37.027977 kernel: ACPI: FACS 0x000000008C66DF80 000040 Nov 8 01:37:37.027983 kernel: ACPI: APIC 0x000000008C58B788 00012C (v04 01072009 AMI 00010013) Nov 8 01:37:37.027989 kernel: ACPI: FPDT 0x000000008C58B8B8 000044 (v01 01072009 AMI 00010013) Nov 8 01:37:37.027994 kernel: ACPI: FIDT 0x000000008C58B900 00009C (v01 SUPERM SMCI--MB 01072009 AMI 00010013) Nov 8 01:37:37.027999 kernel: ACPI: MCFG 0x000000008C58B9A0 00003C (v01 SUPERM SMCI--MB 01072009 MSFT 00000097) Nov 8 01:37:37.028004 kernel: ACPI: SPMI 0x000000008C58B9E0 000041 (v05 SUPERM SMCI--MB 00000000 AMI. 00000000) Nov 8 01:37:37.028009 kernel: ACPI: SSDT 0x000000008C58BA28 001B1C (v02 CpuRef CpuSsdt 00003000 INTL 20160527) Nov 8 01:37:37.028014 kernel: ACPI: SSDT 0x000000008C58D548 0031C6 (v02 SaSsdt SaSsdt 00003000 INTL 20160527) Nov 8 01:37:37.028019 kernel: ACPI: SSDT 0x000000008C590710 00232B (v02 PegSsd PegSsdt 00001000 INTL 20160527) Nov 8 01:37:37.028025 kernel: ACPI: HPET 0x000000008C592A40 000038 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028030 kernel: ACPI: SSDT 0x000000008C592A78 000FAE (v02 SUPERM Ther_Rvp 00001000 INTL 20160527) Nov 8 01:37:37.028035 kernel: ACPI: SSDT 0x000000008C593A28 0008F4 (v02 INTEL xh_mossb 00000000 INTL 20160527) Nov 8 01:37:37.028040 kernel: ACPI: UEFI 0x000000008C594320 000042 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028045 kernel: ACPI: LPIT 0x000000008C594368 000094 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028050 kernel: ACPI: SSDT 0x000000008C594400 0027DE (v02 SUPERM PtidDevc 00001000 INTL 20160527) Nov 8 01:37:37.028056 kernel: ACPI: SSDT 0x000000008C596BE0 0014E2 (v02 SUPERM TbtTypeC 00000000 INTL 20160527) Nov 8 01:37:37.028061 kernel: ACPI: DBGP 0x000000008C5980C8 000034 (v01 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028066 kernel: ACPI: DBG2 0x000000008C598100 000054 (v00 SUPERM SMCI--MB 00000002 01000013) Nov 8 01:37:37.028072 kernel: ACPI: SSDT 0x000000008C598158 001B67 (v02 SUPERM UsbCTabl 00001000 INTL 20160527) Nov 8 01:37:37.028077 kernel: ACPI: DMAR 0x000000008C599CC0 000070 (v01 INTEL EDK2 00000002 01000013) Nov 8 01:37:37.028082 kernel: ACPI: SSDT 0x000000008C599D30 000144 (v02 Intel ADebTabl 00001000 INTL 20160527) Nov 8 01:37:37.028087 kernel: ACPI: TPM2 0x000000008C599E78 000034 (v04 SUPERM SMCI--MB 00000001 AMI 00000000) Nov 8 01:37:37.028092 kernel: ACPI: SSDT 0x000000008C599EB0 000D8F (v02 INTEL SpsNm 00000002 INTL 20160527) Nov 8 01:37:37.028097 kernel: ACPI: WSMT 0x000000008C59AC40 000028 (v01 SUPERM 01072009 AMI 00010013) Nov 8 01:37:37.028102 kernel: ACPI: EINJ 0x000000008C59AC68 000130 (v01 AMI AMI.EINJ 00000000 AMI. 00000000) Nov 8 01:37:37.028107 kernel: ACPI: ERST 0x000000008C59AD98 000230 (v01 AMIER AMI.ERST 00000000 AMI. 00000000) Nov 8 01:37:37.028113 kernel: ACPI: BERT 0x000000008C59AFC8 000030 (v01 AMI AMI.BERT 00000000 AMI. 00000000) Nov 8 01:37:37.028118 kernel: ACPI: HEST 0x000000008C59AFF8 00027C (v01 AMI AMI.HEST 00000000 AMI. 00000000) Nov 8 01:37:37.028123 kernel: ACPI: SSDT 0x000000008C59B278 000162 (v01 SUPERM SMCCDN 00000000 INTL 20181221) Nov 8 01:37:37.028128 kernel: ACPI: Reserving FACP table memory at [mem 0x8c58b670-0x8c58b783] Nov 8 01:37:37.028133 kernel: ACPI: Reserving DSDT table memory at [mem 0x8c54f268-0x8c58b66b] Nov 8 01:37:37.028138 kernel: ACPI: Reserving FACS table memory at [mem 0x8c66df80-0x8c66dfbf] Nov 8 01:37:37.028147 kernel: ACPI: Reserving APIC table memory at [mem 0x8c58b788-0x8c58b8b3] Nov 8 01:37:37.028152 kernel: ACPI: Reserving FPDT table memory at [mem 0x8c58b8b8-0x8c58b8fb] Nov 8 01:37:37.028177 kernel: ACPI: Reserving FIDT table memory at [mem 0x8c58b900-0x8c58b99b] Nov 8 01:37:37.028183 kernel: ACPI: Reserving MCFG table memory at [mem 0x8c58b9a0-0x8c58b9db] Nov 8 01:37:37.028189 kernel: ACPI: Reserving SPMI table memory at [mem 0x8c58b9e0-0x8c58ba20] Nov 8 01:37:37.028194 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58ba28-0x8c58d543] Nov 8 01:37:37.028212 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c58d548-0x8c59070d] Nov 8 01:37:37.028217 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c590710-0x8c592a3a] Nov 8 01:37:37.028222 kernel: ACPI: Reserving HPET table memory at [mem 0x8c592a40-0x8c592a77] Nov 8 01:37:37.028227 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c592a78-0x8c593a25] Nov 8 01:37:37.028232 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c593a28-0x8c59431b] Nov 8 01:37:37.028237 kernel: ACPI: Reserving UEFI table memory at [mem 0x8c594320-0x8c594361] Nov 8 01:37:37.028243 kernel: ACPI: Reserving LPIT table memory at [mem 0x8c594368-0x8c5943fb] Nov 8 01:37:37.028248 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c594400-0x8c596bdd] Nov 8 01:37:37.028253 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c596be0-0x8c5980c1] Nov 8 01:37:37.028258 kernel: ACPI: Reserving DBGP table memory at [mem 0x8c5980c8-0x8c5980fb] Nov 8 01:37:37.028263 kernel: ACPI: Reserving DBG2 table memory at [mem 0x8c598100-0x8c598153] Nov 8 01:37:37.028268 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c598158-0x8c599cbe] Nov 8 01:37:37.028273 kernel: ACPI: Reserving DMAR table memory at [mem 0x8c599cc0-0x8c599d2f] Nov 8 01:37:37.028278 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599d30-0x8c599e73] Nov 8 01:37:37.028283 kernel: ACPI: Reserving TPM2 table memory at [mem 0x8c599e78-0x8c599eab] Nov 8 01:37:37.028289 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c599eb0-0x8c59ac3e] Nov 8 01:37:37.028294 kernel: ACPI: Reserving WSMT table memory at [mem 0x8c59ac40-0x8c59ac67] Nov 8 01:37:37.028299 kernel: ACPI: Reserving EINJ table memory at [mem 0x8c59ac68-0x8c59ad97] Nov 8 01:37:37.028304 kernel: ACPI: Reserving ERST table memory at [mem 0x8c59ad98-0x8c59afc7] Nov 8 01:37:37.028309 kernel: ACPI: Reserving BERT table memory at [mem 0x8c59afc8-0x8c59aff7] Nov 8 01:37:37.028315 kernel: ACPI: Reserving HEST table memory at [mem 0x8c59aff8-0x8c59b273] Nov 8 01:37:37.028320 kernel: ACPI: Reserving SSDT table memory at [mem 0x8c59b278-0x8c59b3d9] Nov 8 01:37:37.028325 kernel: No NUMA configuration found Nov 8 01:37:37.028330 kernel: Faking a node at [mem 0x0000000000000000-0x000000086effffff] Nov 8 01:37:37.028335 kernel: NODE_DATA(0) allocated [mem 0x86effa000-0x86effffff] Nov 8 01:37:37.028341 kernel: Zone ranges: Nov 8 01:37:37.028346 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Nov 8 01:37:37.028351 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Nov 8 01:37:37.028356 kernel: Normal [mem 0x0000000100000000-0x000000086effffff] Nov 8 01:37:37.028362 kernel: Movable zone start for each node Nov 8 01:37:37.028367 kernel: Early memory node ranges Nov 8 01:37:37.028372 kernel: node 0: [mem 0x0000000000001000-0x0000000000098fff] Nov 8 01:37:37.028377 kernel: node 0: [mem 0x0000000000100000-0x000000003fffffff] Nov 8 01:37:37.028382 kernel: node 0: [mem 0x0000000040400000-0x00000000825bdfff] Nov 8 01:37:37.028388 kernel: node 0: [mem 0x00000000825c0000-0x000000008afcdfff] Nov 8 01:37:37.028393 kernel: node 0: [mem 0x000000008c0b3000-0x000000008c23bfff] Nov 8 01:37:37.028398 kernel: node 0: [mem 0x000000008eeff000-0x000000008eefffff] Nov 8 01:37:37.028403 kernel: node 0: [mem 0x0000000100000000-0x000000086effffff] Nov 8 01:37:37.028412 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000086effffff] Nov 8 01:37:37.028418 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Nov 8 01:37:37.028424 kernel: On node 0, zone DMA: 103 pages in unavailable ranges Nov 8 01:37:37.028429 kernel: On node 0, zone DMA32: 1024 pages in unavailable ranges Nov 8 01:37:37.028435 kernel: On node 0, zone DMA32: 2 pages in unavailable ranges Nov 8 01:37:37.028441 kernel: On node 0, zone DMA32: 4325 pages in unavailable ranges Nov 8 01:37:37.028446 kernel: On node 0, zone DMA32: 11459 pages in unavailable ranges Nov 8 01:37:37.028452 kernel: On node 0, zone Normal: 4352 pages in unavailable ranges Nov 8 01:37:37.028458 kernel: On node 0, zone Normal: 4096 pages in unavailable ranges Nov 8 01:37:37.028463 kernel: ACPI: PM-Timer IO Port: 0x1808 Nov 8 01:37:37.028468 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Nov 8 01:37:37.028474 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Nov 8 01:37:37.028479 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Nov 8 01:37:37.028486 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Nov 8 01:37:37.028491 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Nov 8 01:37:37.028496 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Nov 8 01:37:37.028502 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Nov 8 01:37:37.028507 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Nov 8 01:37:37.028513 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Nov 8 01:37:37.028518 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Nov 8 01:37:37.028523 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Nov 8 01:37:37.028529 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Nov 8 01:37:37.028535 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Nov 8 01:37:37.028540 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Nov 8 01:37:37.028546 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Nov 8 01:37:37.028551 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Nov 8 01:37:37.028556 kernel: IOAPIC[0]: apic_id 2, version 32, address 0xfec00000, GSI 0-119 Nov 8 01:37:37.028562 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Nov 8 01:37:37.028567 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Nov 8 01:37:37.028573 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Nov 8 01:37:37.028578 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Nov 8 01:37:37.028585 kernel: TSC deadline timer available Nov 8 01:37:37.028590 kernel: smpboot: Allowing 16 CPUs, 0 hotplug CPUs Nov 8 01:37:37.028596 kernel: [mem 0x90000000-0xdfffffff] available for PCI devices Nov 8 01:37:37.028601 kernel: Booting paravirtualized kernel on bare hardware Nov 8 01:37:37.028607 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Nov 8 01:37:37.028612 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Nov 8 01:37:37.028618 kernel: percpu: Embedded 58 pages/cpu s196712 r8192 d32664 u262144 Nov 8 01:37:37.028623 kernel: pcpu-alloc: s196712 r8192 d32664 u262144 alloc=1*2097152 Nov 8 01:37:37.028628 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Nov 8 01:37:37.028635 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=480a02cf7f2001774aa495c3e719d4173e968e6839485a7d2b207ef2facca472 Nov 8 01:37:37.028641 kernel: random: crng init done Nov 8 01:37:37.028646 kernel: Dentry cache hash table entries: 4194304 (order: 13, 33554432 bytes, linear) Nov 8 01:37:37.028652 kernel: Inode-cache hash table entries: 2097152 (order: 12, 16777216 bytes, linear) Nov 8 01:37:37.028657 kernel: Fallback order for Node 0: 0 Nov 8 01:37:37.028663 kernel: Built 1 zonelists, mobility grouping on. Total pages: 8232416 Nov 8 01:37:37.028668 kernel: Policy zone: Normal Nov 8 01:37:37.028674 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 8 01:37:37.028680 kernel: software IO TLB: area num 16. Nov 8 01:37:37.028686 kernel: Memory: 32720312K/33452984K available (12288K kernel code, 2288K rwdata, 22748K rodata, 42880K init, 2320K bss, 732412K reserved, 0K cma-reserved) Nov 8 01:37:37.028692 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Nov 8 01:37:37.028697 kernel: ftrace: allocating 37980 entries in 149 pages Nov 8 01:37:37.028702 kernel: ftrace: allocated 149 pages with 4 groups Nov 8 01:37:37.028708 kernel: Dynamic Preempt: voluntary Nov 8 01:37:37.028714 kernel: rcu: Preemptible hierarchical RCU implementation. Nov 8 01:37:37.028719 kernel: rcu: RCU event tracing is enabled. Nov 8 01:37:37.028725 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Nov 8 01:37:37.028732 kernel: Trampoline variant of Tasks RCU enabled. Nov 8 01:37:37.028738 kernel: Rude variant of Tasks RCU enabled. Nov 8 01:37:37.028743 kernel: Tracing variant of Tasks RCU enabled. Nov 8 01:37:37.028748 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 8 01:37:37.028754 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Nov 8 01:37:37.028759 kernel: NR_IRQS: 33024, nr_irqs: 2184, preallocated irqs: 16 Nov 8 01:37:37.028765 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 8 01:37:37.028770 kernel: Console: colour dummy device 80x25 Nov 8 01:37:37.028776 kernel: printk: console [tty0] enabled Nov 8 01:37:37.028781 kernel: printk: console [ttyS1] enabled Nov 8 01:37:37.028788 kernel: ACPI: Core revision 20230628 Nov 8 01:37:37.028793 kernel: hpet: HPET dysfunctional in PC10. Force disabled. Nov 8 01:37:37.028799 kernel: APIC: Switch to symmetric I/O mode setup Nov 8 01:37:37.028804 kernel: DMAR: Host address width 39 Nov 8 01:37:37.028810 kernel: DMAR: DRHD base: 0x000000fed91000 flags: 0x1 Nov 8 01:37:37.028815 kernel: DMAR: dmar0: reg_base_addr fed91000 ver 1:0 cap d2008c40660462 ecap f050da Nov 8 01:37:37.028821 kernel: DMAR: RMRR base: 0x0000008cf19000 end: 0x0000008d162fff Nov 8 01:37:37.028826 kernel: DMAR-IR: IOAPIC id 2 under DRHD base 0xfed91000 IOMMU 0 Nov 8 01:37:37.028832 kernel: DMAR-IR: HPET id 0 under DRHD base 0xfed91000 Nov 8 01:37:37.028838 kernel: DMAR-IR: Queued invalidation will be enabled to support x2apic and Intr-remapping. Nov 8 01:37:37.028843 kernel: DMAR-IR: Enabled IRQ remapping in x2apic mode Nov 8 01:37:37.028849 kernel: x2apic enabled Nov 8 01:37:37.028855 kernel: APIC: Switched APIC routing to: cluster x2apic Nov 8 01:37:37.028860 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x3101f59f5e6, max_idle_ns: 440795259996 ns Nov 8 01:37:37.028866 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 6799.81 BogoMIPS (lpj=3399906) Nov 8 01:37:37.028871 kernel: CPU0: Thermal monitoring enabled (TM1) Nov 8 01:37:37.028877 kernel: process: using mwait in idle threads Nov 8 01:37:37.028882 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Nov 8 01:37:37.028889 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Nov 8 01:37:37.028894 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Nov 8 01:37:37.028899 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Nov 8 01:37:37.028905 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Nov 8 01:37:37.028910 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Nov 8 01:37:37.028915 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Nov 8 01:37:37.028921 kernel: RETBleed: Mitigation: Enhanced IBRS Nov 8 01:37:37.028926 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Nov 8 01:37:37.028931 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Nov 8 01:37:37.028937 kernel: TAA: Mitigation: TSX disabled Nov 8 01:37:37.028942 kernel: MMIO Stale Data: Mitigation: Clear CPU buffers Nov 8 01:37:37.028948 kernel: SRBDS: Mitigation: Microcode Nov 8 01:37:37.028954 kernel: GDS: Mitigation: Microcode Nov 8 01:37:37.028959 kernel: active return thunk: its_return_thunk Nov 8 01:37:37.028965 kernel: ITS: Mitigation: Aligned branch/return thunks Nov 8 01:37:37.028970 kernel: VMSCAPE: Mitigation: IBPB before exit to userspace Nov 8 01:37:37.028975 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Nov 8 01:37:37.028981 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Nov 8 01:37:37.028986 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Nov 8 01:37:37.028992 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Nov 8 01:37:37.028997 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Nov 8 01:37:37.029002 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Nov 8 01:37:37.029009 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Nov 8 01:37:37.029014 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Nov 8 01:37:37.029019 kernel: x86/fpu: Enabled xstate features 0x1f, context size is 960 bytes, using 'compacted' format. Nov 8 01:37:37.029025 kernel: Freeing SMP alternatives memory: 32K Nov 8 01:37:37.029030 kernel: pid_max: default: 32768 minimum: 301 Nov 8 01:37:37.029036 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Nov 8 01:37:37.029041 kernel: landlock: Up and running. Nov 8 01:37:37.029046 kernel: SELinux: Initializing. Nov 8 01:37:37.029052 kernel: Mount-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.029057 kernel: Mountpoint-cache hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.029063 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Nov 8 01:37:37.029068 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Nov 8 01:37:37.029074 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Nov 8 01:37:37.029080 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Nov 8 01:37:37.029086 kernel: Performance Events: PEBS fmt3+, Skylake events, 32-deep LBR, full-width counters, Intel PMU driver. Nov 8 01:37:37.029091 kernel: ... version: 4 Nov 8 01:37:37.029097 kernel: ... bit width: 48 Nov 8 01:37:37.029102 kernel: ... generic registers: 4 Nov 8 01:37:37.029107 kernel: ... value mask: 0000ffffffffffff Nov 8 01:37:37.029113 kernel: ... max period: 00007fffffffffff Nov 8 01:37:37.029118 kernel: ... fixed-purpose events: 3 Nov 8 01:37:37.029125 kernel: ... event mask: 000000070000000f Nov 8 01:37:37.029130 kernel: signal: max sigframe size: 2032 Nov 8 01:37:37.029136 kernel: Estimated ratio of average max frequency by base frequency (times 1024): 1445 Nov 8 01:37:37.029147 kernel: rcu: Hierarchical SRCU implementation. Nov 8 01:37:37.029153 kernel: rcu: Max phase no-delay instances is 400. Nov 8 01:37:37.029179 kernel: NMI watchdog: Enabled. Permanently consumes one hw-PMU counter. Nov 8 01:37:37.029185 kernel: smp: Bringing up secondary CPUs ... Nov 8 01:37:37.029191 kernel: smpboot: x86: Booting SMP configuration: Nov 8 01:37:37.029211 kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6 #7 #8 #9 #10 #11 #12 #13 #14 #15 Nov 8 01:37:37.029218 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Nov 8 01:37:37.029224 kernel: smp: Brought up 1 node, 16 CPUs Nov 8 01:37:37.029229 kernel: smpboot: Max logical packages: 1 Nov 8 01:37:37.029234 kernel: smpboot: Total of 16 processors activated (108796.99 BogoMIPS) Nov 8 01:37:37.029240 kernel: devtmpfs: initialized Nov 8 01:37:37.029245 kernel: x86/mm: Memory block size: 128MB Nov 8 01:37:37.029251 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x825be000-0x825befff] (4096 bytes) Nov 8 01:37:37.029256 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x8c23c000-0x8c66dfff] (4399104 bytes) Nov 8 01:37:37.029263 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 8 01:37:37.029268 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Nov 8 01:37:37.029274 kernel: pinctrl core: initialized pinctrl subsystem Nov 8 01:37:37.029279 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 8 01:37:37.029284 kernel: audit: initializing netlink subsys (disabled) Nov 8 01:37:37.029290 kernel: audit: type=2000 audit(1762565851.039:1): state=initialized audit_enabled=0 res=1 Nov 8 01:37:37.029295 kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 8 01:37:37.029301 kernel: thermal_sys: Registered thermal governor 'user_space' Nov 8 01:37:37.029306 kernel: cpuidle: using governor menu Nov 8 01:37:37.029313 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 8 01:37:37.029318 kernel: dca service started, version 1.12.1 Nov 8 01:37:37.029324 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Nov 8 01:37:37.029329 kernel: PCI: Using configuration type 1 for base access Nov 8 01:37:37.029335 kernel: ENERGY_PERF_BIAS: Set to 'normal', was 'performance' Nov 8 01:37:37.029340 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Nov 8 01:37:37.029345 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Nov 8 01:37:37.029351 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Nov 8 01:37:37.029356 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Nov 8 01:37:37.029363 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Nov 8 01:37:37.029368 kernel: ACPI: Added _OSI(Module Device) Nov 8 01:37:37.029374 kernel: ACPI: Added _OSI(Processor Device) Nov 8 01:37:37.029379 kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 8 01:37:37.029384 kernel: ACPI: 12 ACPI AML tables successfully acquired and loaded Nov 8 01:37:37.029390 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029395 kernel: ACPI: SSDT 0xFFFF9E9CC1B36800 000400 (v02 PmRef Cpu0Cst 00003001 INTL 20160527) Nov 8 01:37:37.029401 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029406 kernel: ACPI: SSDT 0xFFFF9E9CC1B2A000 000683 (v02 PmRef Cpu0Ist 00003000 INTL 20160527) Nov 8 01:37:37.029413 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029418 kernel: ACPI: SSDT 0xFFFF9E9CC0247400 0000F4 (v02 PmRef Cpu0Psd 00003000 INTL 20160527) Nov 8 01:37:37.029424 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029429 kernel: ACPI: SSDT 0xFFFF9E9CC1E5C800 0005FC (v02 PmRef ApIst 00003000 INTL 20160527) Nov 8 01:37:37.029434 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029439 kernel: ACPI: SSDT 0xFFFF9E9CC012F000 000AB0 (v02 PmRef ApPsd 00003000 INTL 20160527) Nov 8 01:37:37.029445 kernel: ACPI: Dynamic OEM Table Load: Nov 8 01:37:37.029450 kernel: ACPI: SSDT 0xFFFF9E9CC1B30400 00030A (v02 PmRef ApCst 00003000 INTL 20160527) Nov 8 01:37:37.029456 kernel: ACPI: _OSC evaluated successfully for all CPUs Nov 8 01:37:37.029461 kernel: ACPI: Interpreter enabled Nov 8 01:37:37.029468 kernel: ACPI: PM: (supports S0 S5) Nov 8 01:37:37.029473 kernel: ACPI: Using IOAPIC for interrupt routing Nov 8 01:37:37.029478 kernel: HEST: Enabling Firmware First mode for corrected errors. Nov 8 01:37:37.029484 kernel: mce: [Firmware Bug]: Ignoring request to disable invalid MCA bank 14. Nov 8 01:37:37.029489 kernel: HEST: Table parsing has been initialized. Nov 8 01:37:37.029495 kernel: GHES: APEI firmware first mode is enabled by APEI bit and WHEA _OSC. Nov 8 01:37:37.029500 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Nov 8 01:37:37.029506 kernel: PCI: Ignoring E820 reservations for host bridge windows Nov 8 01:37:37.029511 kernel: ACPI: Enabled 9 GPEs in block 00 to 7F Nov 8 01:37:37.029518 kernel: ACPI: \_SB_.PCI0.XDCI.USBC: New power resource Nov 8 01:37:37.029523 kernel: ACPI: \_SB_.PCI0.SAT0.VOL0.V0PR: New power resource Nov 8 01:37:37.029529 kernel: ACPI: \_SB_.PCI0.SAT0.VOL1.V1PR: New power resource Nov 8 01:37:37.029534 kernel: ACPI: \_SB_.PCI0.SAT0.VOL2.V2PR: New power resource Nov 8 01:37:37.029540 kernel: ACPI: \_SB_.PCI0.CNVW.WRST: New power resource Nov 8 01:37:37.029545 kernel: ACPI: \_TZ_.FN00: New power resource Nov 8 01:37:37.029551 kernel: ACPI: \_TZ_.FN01: New power resource Nov 8 01:37:37.029556 kernel: ACPI: \_TZ_.FN02: New power resource Nov 8 01:37:37.029561 kernel: ACPI: \_TZ_.FN03: New power resource Nov 8 01:37:37.029568 kernel: ACPI: \_TZ_.FN04: New power resource Nov 8 01:37:37.029573 kernel: ACPI: \PIN_: New power resource Nov 8 01:37:37.029579 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-fe]) Nov 8 01:37:37.029654 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Nov 8 01:37:37.029708 kernel: acpi PNP0A08:00: _OSC: platform does not support [AER] Nov 8 01:37:37.029758 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability LTR] Nov 8 01:37:37.029766 kernel: PCI host bridge to bus 0000:00 Nov 8 01:37:37.029817 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Nov 8 01:37:37.029862 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Nov 8 01:37:37.029905 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Nov 8 01:37:37.029948 kernel: pci_bus 0000:00: root bus resource [mem 0x90000000-0xdfffffff window] Nov 8 01:37:37.029989 kernel: pci_bus 0000:00: root bus resource [mem 0xfc800000-0xfe7fffff window] Nov 8 01:37:37.030032 kernel: pci_bus 0000:00: root bus resource [bus 00-fe] Nov 8 01:37:37.030090 kernel: pci 0000:00:00.0: [8086:3e31] type 00 class 0x060000 Nov 8 01:37:37.030153 kernel: pci 0000:00:01.0: [8086:1901] type 01 class 0x060400 Nov 8 01:37:37.030239 kernel: pci 0000:00:01.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.030291 kernel: pci 0000:00:08.0: [8086:1911] type 00 class 0x088000 Nov 8 01:37:37.030341 kernel: pci 0000:00:08.0: reg 0x10: [mem 0x9551f000-0x9551ffff 64bit] Nov 8 01:37:37.030393 kernel: pci 0000:00:12.0: [8086:a379] type 00 class 0x118000 Nov 8 01:37:37.030442 kernel: pci 0000:00:12.0: reg 0x10: [mem 0x9551e000-0x9551efff 64bit] Nov 8 01:37:37.030497 kernel: pci 0000:00:14.0: [8086:a36d] type 00 class 0x0c0330 Nov 8 01:37:37.030546 kernel: pci 0000:00:14.0: reg 0x10: [mem 0x95500000-0x9550ffff 64bit] Nov 8 01:37:37.030593 kernel: pci 0000:00:14.0: PME# supported from D3hot D3cold Nov 8 01:37:37.030646 kernel: pci 0000:00:14.2: [8086:a36f] type 00 class 0x050000 Nov 8 01:37:37.030695 kernel: pci 0000:00:14.2: reg 0x10: [mem 0x95512000-0x95513fff 64bit] Nov 8 01:37:37.030744 kernel: pci 0000:00:14.2: reg 0x18: [mem 0x9551d000-0x9551dfff 64bit] Nov 8 01:37:37.030798 kernel: pci 0000:00:15.0: [8086:a368] type 00 class 0x0c8000 Nov 8 01:37:37.030847 kernel: pci 0000:00:15.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Nov 8 01:37:37.030902 kernel: pci 0000:00:15.1: [8086:a369] type 00 class 0x0c8000 Nov 8 01:37:37.030951 kernel: pci 0000:00:15.1: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Nov 8 01:37:37.031003 kernel: pci 0000:00:16.0: [8086:a360] type 00 class 0x078000 Nov 8 01:37:37.031051 kernel: pci 0000:00:16.0: reg 0x10: [mem 0x9551a000-0x9551afff 64bit] Nov 8 01:37:37.031102 kernel: pci 0000:00:16.0: PME# supported from D3hot Nov 8 01:37:37.031174 kernel: pci 0000:00:16.1: [8086:a361] type 00 class 0x078000 Nov 8 01:37:37.031246 kernel: pci 0000:00:16.1: reg 0x10: [mem 0x95519000-0x95519fff 64bit] Nov 8 01:37:37.031295 kernel: pci 0000:00:16.1: PME# supported from D3hot Nov 8 01:37:37.031348 kernel: pci 0000:00:16.4: [8086:a364] type 00 class 0x078000 Nov 8 01:37:37.031396 kernel: pci 0000:00:16.4: reg 0x10: [mem 0x95518000-0x95518fff 64bit] Nov 8 01:37:37.031445 kernel: pci 0000:00:16.4: PME# supported from D3hot Nov 8 01:37:37.031500 kernel: pci 0000:00:17.0: [8086:a352] type 00 class 0x010601 Nov 8 01:37:37.031551 kernel: pci 0000:00:17.0: reg 0x10: [mem 0x95510000-0x95511fff] Nov 8 01:37:37.031599 kernel: pci 0000:00:17.0: reg 0x14: [mem 0x95517000-0x955170ff] Nov 8 01:37:37.031648 kernel: pci 0000:00:17.0: reg 0x18: [io 0x6050-0x6057] Nov 8 01:37:37.031695 kernel: pci 0000:00:17.0: reg 0x1c: [io 0x6040-0x6043] Nov 8 01:37:37.031743 kernel: pci 0000:00:17.0: reg 0x20: [io 0x6020-0x603f] Nov 8 01:37:37.031794 kernel: pci 0000:00:17.0: reg 0x24: [mem 0x95516000-0x955167ff] Nov 8 01:37:37.031842 kernel: pci 0000:00:17.0: PME# supported from D3hot Nov 8 01:37:37.031895 kernel: pci 0000:00:1b.0: [8086:a340] type 01 class 0x060400 Nov 8 01:37:37.031945 kernel: pci 0000:00:1b.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032001 kernel: pci 0000:00:1b.4: [8086:a32c] type 01 class 0x060400 Nov 8 01:37:37.032053 kernel: pci 0000:00:1b.4: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032106 kernel: pci 0000:00:1b.5: [8086:a32d] type 01 class 0x060400 Nov 8 01:37:37.032174 kernel: pci 0000:00:1b.5: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032242 kernel: pci 0000:00:1c.0: [8086:a338] type 01 class 0x060400 Nov 8 01:37:37.032291 kernel: pci 0000:00:1c.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032345 kernel: pci 0000:00:1c.3: [8086:a33b] type 01 class 0x060400 Nov 8 01:37:37.032397 kernel: pci 0000:00:1c.3: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.032451 kernel: pci 0000:00:1e.0: [8086:a328] type 00 class 0x078000 Nov 8 01:37:37.032501 kernel: pci 0000:00:1e.0: reg 0x10: [mem 0x00000000-0x00000fff 64bit] Nov 8 01:37:37.032554 kernel: pci 0000:00:1f.0: [8086:a309] type 00 class 0x060100 Nov 8 01:37:37.032608 kernel: pci 0000:00:1f.4: [8086:a323] type 00 class 0x0c0500 Nov 8 01:37:37.032658 kernel: pci 0000:00:1f.4: reg 0x10: [mem 0x95514000-0x955140ff 64bit] Nov 8 01:37:37.032709 kernel: pci 0000:00:1f.4: reg 0x20: [io 0xefa0-0xefbf] Nov 8 01:37:37.032761 kernel: pci 0000:00:1f.5: [8086:a324] type 00 class 0x0c8000 Nov 8 01:37:37.032811 kernel: pci 0000:00:1f.5: reg 0x10: [mem 0xfe010000-0xfe010fff] Nov 8 01:37:37.032865 kernel: pci 0000:01:00.0: [15b3:1015] type 00 class 0x020000 Nov 8 01:37:37.032917 kernel: pci 0000:01:00.0: reg 0x10: [mem 0x92000000-0x93ffffff 64bit pref] Nov 8 01:37:37.032968 kernel: pci 0000:01:00.0: reg 0x30: [mem 0x95200000-0x952fffff pref] Nov 8 01:37:37.033021 kernel: pci 0000:01:00.0: PME# supported from D3cold Nov 8 01:37:37.033071 kernel: pci 0000:01:00.0: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Nov 8 01:37:37.033122 kernel: pci 0000:01:00.0: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Nov 8 01:37:37.033214 kernel: pci 0000:01:00.1: [15b3:1015] type 00 class 0x020000 Nov 8 01:37:37.033265 kernel: pci 0000:01:00.1: reg 0x10: [mem 0x90000000-0x91ffffff 64bit pref] Nov 8 01:37:37.033317 kernel: pci 0000:01:00.1: reg 0x30: [mem 0x95100000-0x951fffff pref] Nov 8 01:37:37.033367 kernel: pci 0000:01:00.1: PME# supported from D3cold Nov 8 01:37:37.033420 kernel: pci 0000:01:00.1: reg 0x1a4: [mem 0x00000000-0x000fffff 64bit pref] Nov 8 01:37:37.033470 kernel: pci 0000:01:00.1: VF(n) BAR0 space: [mem 0x00000000-0x007fffff 64bit pref] (contains BAR0 for 8 VFs) Nov 8 01:37:37.033520 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Nov 8 01:37:37.033569 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Nov 8 01:37:37.033620 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Nov 8 01:37:37.033670 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Nov 8 01:37:37.033725 kernel: pci 0000:03:00.0: working around ROM BAR overlap defect Nov 8 01:37:37.033777 kernel: pci 0000:03:00.0: [8086:1533] type 00 class 0x020000 Nov 8 01:37:37.033829 kernel: pci 0000:03:00.0: reg 0x10: [mem 0x95400000-0x9547ffff] Nov 8 01:37:37.033880 kernel: pci 0000:03:00.0: reg 0x18: [io 0x5000-0x501f] Nov 8 01:37:37.033930 kernel: pci 0000:03:00.0: reg 0x1c: [mem 0x95480000-0x95483fff] Nov 8 01:37:37.033980 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.034029 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Nov 8 01:37:37.034079 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Nov 8 01:37:37.034127 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Nov 8 01:37:37.034224 kernel: pci 0000:04:00.0: working around ROM BAR overlap defect Nov 8 01:37:37.034277 kernel: pci 0000:04:00.0: [8086:1533] type 00 class 0x020000 Nov 8 01:37:37.034327 kernel: pci 0000:04:00.0: reg 0x10: [mem 0x95300000-0x9537ffff] Nov 8 01:37:37.034378 kernel: pci 0000:04:00.0: reg 0x18: [io 0x4000-0x401f] Nov 8 01:37:37.034427 kernel: pci 0000:04:00.0: reg 0x1c: [mem 0x95380000-0x95383fff] Nov 8 01:37:37.034478 kernel: pci 0000:04:00.0: PME# supported from D0 D3hot D3cold Nov 8 01:37:37.034527 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Nov 8 01:37:37.034576 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Nov 8 01:37:37.034628 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Nov 8 01:37:37.034678 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Nov 8 01:37:37.034733 kernel: pci 0000:06:00.0: [1a03:1150] type 01 class 0x060400 Nov 8 01:37:37.034787 kernel: pci 0000:06:00.0: enabling Extended Tags Nov 8 01:37:37.034837 kernel: pci 0000:06:00.0: supports D1 D2 Nov 8 01:37:37.034887 kernel: pci 0000:06:00.0: PME# supported from D0 D1 D2 D3hot D3cold Nov 8 01:37:37.034937 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Nov 8 01:37:37.034988 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.035039 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.035095 kernel: pci_bus 0000:07: extended config space not accessible Nov 8 01:37:37.035155 kernel: pci 0000:07:00.0: [1a03:2000] type 00 class 0x030000 Nov 8 01:37:37.035255 kernel: pci 0000:07:00.0: reg 0x10: [mem 0x94000000-0x94ffffff] Nov 8 01:37:37.035307 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x95000000-0x9501ffff] Nov 8 01:37:37.035359 kernel: pci 0000:07:00.0: reg 0x18: [io 0x3000-0x307f] Nov 8 01:37:37.035413 kernel: pci 0000:07:00.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Nov 8 01:37:37.035465 kernel: pci 0000:07:00.0: supports D1 D2 Nov 8 01:37:37.035518 kernel: pci 0000:07:00.0: PME# supported from D0 D1 D2 D3hot D3cold Nov 8 01:37:37.035568 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Nov 8 01:37:37.035618 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.035667 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.035675 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 0 Nov 8 01:37:37.035681 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 1 Nov 8 01:37:37.035689 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 0 Nov 8 01:37:37.035694 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 0 Nov 8 01:37:37.035700 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 0 Nov 8 01:37:37.035706 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 0 Nov 8 01:37:37.035712 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 0 Nov 8 01:37:37.035717 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 0 Nov 8 01:37:37.035723 kernel: iommu: Default domain type: Translated Nov 8 01:37:37.035729 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Nov 8 01:37:37.035735 kernel: PCI: Using ACPI for IRQ routing Nov 8 01:37:37.035741 kernel: PCI: pci_cache_line_size set to 64 bytes Nov 8 01:37:37.035747 kernel: e820: reserve RAM buffer [mem 0x00099800-0x0009ffff] Nov 8 01:37:37.035753 kernel: e820: reserve RAM buffer [mem 0x825be000-0x83ffffff] Nov 8 01:37:37.035758 kernel: e820: reserve RAM buffer [mem 0x8afce000-0x8bffffff] Nov 8 01:37:37.035764 kernel: e820: reserve RAM buffer [mem 0x8c23c000-0x8fffffff] Nov 8 01:37:37.035770 kernel: e820: reserve RAM buffer [mem 0x8ef00000-0x8fffffff] Nov 8 01:37:37.035775 kernel: e820: reserve RAM buffer [mem 0x86f000000-0x86fffffff] Nov 8 01:37:37.035826 kernel: pci 0000:07:00.0: vgaarb: setting as boot VGA device Nov 8 01:37:37.035878 kernel: pci 0000:07:00.0: vgaarb: bridge control possible Nov 8 01:37:37.035932 kernel: pci 0000:07:00.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Nov 8 01:37:37.035941 kernel: vgaarb: loaded Nov 8 01:37:37.035947 kernel: clocksource: Switched to clocksource tsc-early Nov 8 01:37:37.035952 kernel: VFS: Disk quotas dquot_6.6.0 Nov 8 01:37:37.035958 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 8 01:37:37.035964 kernel: pnp: PnP ACPI init Nov 8 01:37:37.036014 kernel: system 00:00: [mem 0x40000000-0x403fffff] has been reserved Nov 8 01:37:37.036062 kernel: pnp 00:02: [dma 0 disabled] Nov 8 01:37:37.036115 kernel: pnp 00:03: [dma 0 disabled] Nov 8 01:37:37.036194 kernel: system 00:04: [io 0x0680-0x069f] has been reserved Nov 8 01:37:37.036260 kernel: system 00:04: [io 0x164e-0x164f] has been reserved Nov 8 01:37:37.036307 kernel: system 00:05: [mem 0xfed10000-0xfed17fff] has been reserved Nov 8 01:37:37.036352 kernel: system 00:05: [mem 0xfed18000-0xfed18fff] has been reserved Nov 8 01:37:37.036396 kernel: system 00:05: [mem 0xfed19000-0xfed19fff] has been reserved Nov 8 01:37:37.036444 kernel: system 00:05: [mem 0xe0000000-0xefffffff] has been reserved Nov 8 01:37:37.036488 kernel: system 00:05: [mem 0xfed20000-0xfed3ffff] has been reserved Nov 8 01:37:37.036535 kernel: system 00:05: [mem 0xfed90000-0xfed93fff] could not be reserved Nov 8 01:37:37.036580 kernel: system 00:05: [mem 0xfed45000-0xfed8ffff] has been reserved Nov 8 01:37:37.036625 kernel: system 00:05: [mem 0xfee00000-0xfeefffff] could not be reserved Nov 8 01:37:37.036674 kernel: system 00:06: [io 0x1800-0x18fe] could not be reserved Nov 8 01:37:37.036720 kernel: system 00:06: [mem 0xfd000000-0xfd69ffff] has been reserved Nov 8 01:37:37.036767 kernel: system 00:06: [mem 0xfd6c0000-0xfd6cffff] has been reserved Nov 8 01:37:37.036811 kernel: system 00:06: [mem 0xfd6f0000-0xfdffffff] has been reserved Nov 8 01:37:37.036856 kernel: system 00:06: [mem 0xfe000000-0xfe01ffff] could not be reserved Nov 8 01:37:37.036899 kernel: system 00:06: [mem 0xfe200000-0xfe7fffff] has been reserved Nov 8 01:37:37.036944 kernel: system 00:06: [mem 0xff000000-0xffffffff] has been reserved Nov 8 01:37:37.036991 kernel: system 00:07: [io 0x2000-0x20fe] has been reserved Nov 8 01:37:37.037000 kernel: pnp: PnP ACPI: found 9 devices Nov 8 01:37:37.037008 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Nov 8 01:37:37.037014 kernel: NET: Registered PF_INET protocol family Nov 8 01:37:37.037019 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 8 01:37:37.037025 kernel: tcp_listen_portaddr_hash hash table entries: 16384 (order: 6, 262144 bytes, linear) Nov 8 01:37:37.037031 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 8 01:37:37.037037 kernel: TCP established hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 8 01:37:37.037043 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Nov 8 01:37:37.037050 kernel: TCP: Hash tables configured (established 262144 bind 65536) Nov 8 01:37:37.037055 kernel: UDP hash table entries: 16384 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.037062 kernel: UDP-Lite hash table entries: 16384 (order: 7, 524288 bytes, linear) Nov 8 01:37:37.037068 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 8 01:37:37.037074 kernel: NET: Registered PF_XDP protocol family Nov 8 01:37:37.037124 kernel: pci 0000:00:15.0: BAR 0: assigned [mem 0x95515000-0x95515fff 64bit] Nov 8 01:37:37.037218 kernel: pci 0000:00:15.1: BAR 0: assigned [mem 0x9551b000-0x9551bfff 64bit] Nov 8 01:37:37.037269 kernel: pci 0000:00:1e.0: BAR 0: assigned [mem 0x9551c000-0x9551cfff 64bit] Nov 8 01:37:37.037320 kernel: pci 0000:01:00.0: BAR 7: no space for [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037371 kernel: pci 0000:01:00.0: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037423 kernel: pci 0000:01:00.1: BAR 7: no space for [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037474 kernel: pci 0000:01:00.1: BAR 7: failed to assign [mem size 0x00800000 64bit pref] Nov 8 01:37:37.037523 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Nov 8 01:37:37.037572 kernel: pci 0000:00:01.0: bridge window [mem 0x95100000-0x952fffff] Nov 8 01:37:37.037622 kernel: pci 0000:00:01.0: bridge window [mem 0x90000000-0x93ffffff 64bit pref] Nov 8 01:37:37.037670 kernel: pci 0000:00:1b.0: PCI bridge to [bus 02] Nov 8 01:37:37.037721 kernel: pci 0000:00:1b.4: PCI bridge to [bus 03] Nov 8 01:37:37.037771 kernel: pci 0000:00:1b.4: bridge window [io 0x5000-0x5fff] Nov 8 01:37:37.037821 kernel: pci 0000:00:1b.4: bridge window [mem 0x95400000-0x954fffff] Nov 8 01:37:37.037869 kernel: pci 0000:00:1b.5: PCI bridge to [bus 04] Nov 8 01:37:37.037918 kernel: pci 0000:00:1b.5: bridge window [io 0x4000-0x4fff] Nov 8 01:37:37.037966 kernel: pci 0000:00:1b.5: bridge window [mem 0x95300000-0x953fffff] Nov 8 01:37:37.038015 kernel: pci 0000:00:1c.0: PCI bridge to [bus 05] Nov 8 01:37:37.038067 kernel: pci 0000:06:00.0: PCI bridge to [bus 07] Nov 8 01:37:37.038118 kernel: pci 0000:06:00.0: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.038210 kernel: pci 0000:06:00.0: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.038259 kernel: pci 0000:00:1c.3: PCI bridge to [bus 06-07] Nov 8 01:37:37.038308 kernel: pci 0000:00:1c.3: bridge window [io 0x3000-0x3fff] Nov 8 01:37:37.038356 kernel: pci 0000:00:1c.3: bridge window [mem 0x94000000-0x950fffff] Nov 8 01:37:37.038402 kernel: pci_bus 0000:00: Some PCI device resources are unassigned, try booting with pci=realloc Nov 8 01:37:37.038445 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Nov 8 01:37:37.038491 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Nov 8 01:37:37.038534 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Nov 8 01:37:37.038577 kernel: pci_bus 0000:00: resource 7 [mem 0x90000000-0xdfffffff window] Nov 8 01:37:37.038619 kernel: pci_bus 0000:00: resource 8 [mem 0xfc800000-0xfe7fffff window] Nov 8 01:37:37.038670 kernel: pci_bus 0000:01: resource 1 [mem 0x95100000-0x952fffff] Nov 8 01:37:37.038715 kernel: pci_bus 0000:01: resource 2 [mem 0x90000000-0x93ffffff 64bit pref] Nov 8 01:37:37.038765 kernel: pci_bus 0000:03: resource 0 [io 0x5000-0x5fff] Nov 8 01:37:37.038812 kernel: pci_bus 0000:03: resource 1 [mem 0x95400000-0x954fffff] Nov 8 01:37:37.038865 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Nov 8 01:37:37.038910 kernel: pci_bus 0000:04: resource 1 [mem 0x95300000-0x953fffff] Nov 8 01:37:37.038959 kernel: pci_bus 0000:06: resource 0 [io 0x3000-0x3fff] Nov 8 01:37:37.039004 kernel: pci_bus 0000:06: resource 1 [mem 0x94000000-0x950fffff] Nov 8 01:37:37.039051 kernel: pci_bus 0000:07: resource 0 [io 0x3000-0x3fff] Nov 8 01:37:37.039097 kernel: pci_bus 0000:07: resource 1 [mem 0x94000000-0x950fffff] Nov 8 01:37:37.039106 kernel: PCI: CLS 64 bytes, default 64 Nov 8 01:37:37.039113 kernel: DMAR: No ATSR found Nov 8 01:37:37.039118 kernel: DMAR: No SATC found Nov 8 01:37:37.039124 kernel: DMAR: dmar0: Using Queued invalidation Nov 8 01:37:37.039219 kernel: pci 0000:00:00.0: Adding to iommu group 0 Nov 8 01:37:37.039269 kernel: pci 0000:00:01.0: Adding to iommu group 1 Nov 8 01:37:37.039317 kernel: pci 0000:00:08.0: Adding to iommu group 2 Nov 8 01:37:37.039367 kernel: pci 0000:00:12.0: Adding to iommu group 3 Nov 8 01:37:37.039418 kernel: pci 0000:00:14.0: Adding to iommu group 4 Nov 8 01:37:37.039468 kernel: pci 0000:00:14.2: Adding to iommu group 4 Nov 8 01:37:37.039515 kernel: pci 0000:00:15.0: Adding to iommu group 5 Nov 8 01:37:37.039564 kernel: pci 0000:00:15.1: Adding to iommu group 5 Nov 8 01:37:37.039612 kernel: pci 0000:00:16.0: Adding to iommu group 6 Nov 8 01:37:37.039662 kernel: pci 0000:00:16.1: Adding to iommu group 6 Nov 8 01:37:37.039710 kernel: pci 0000:00:16.4: Adding to iommu group 6 Nov 8 01:37:37.039760 kernel: pci 0000:00:17.0: Adding to iommu group 7 Nov 8 01:37:37.039811 kernel: pci 0000:00:1b.0: Adding to iommu group 8 Nov 8 01:37:37.039861 kernel: pci 0000:00:1b.4: Adding to iommu group 9 Nov 8 01:37:37.039909 kernel: pci 0000:00:1b.5: Adding to iommu group 10 Nov 8 01:37:37.039958 kernel: pci 0000:00:1c.0: Adding to iommu group 11 Nov 8 01:37:37.040007 kernel: pci 0000:00:1c.3: Adding to iommu group 12 Nov 8 01:37:37.040055 kernel: pci 0000:00:1e.0: Adding to iommu group 13 Nov 8 01:37:37.040103 kernel: pci 0000:00:1f.0: Adding to iommu group 14 Nov 8 01:37:37.040177 kernel: pci 0000:00:1f.4: Adding to iommu group 14 Nov 8 01:37:37.040246 kernel: pci 0000:00:1f.5: Adding to iommu group 14 Nov 8 01:37:37.040299 kernel: pci 0000:01:00.0: Adding to iommu group 1 Nov 8 01:37:37.040350 kernel: pci 0000:01:00.1: Adding to iommu group 1 Nov 8 01:37:37.040400 kernel: pci 0000:03:00.0: Adding to iommu group 15 Nov 8 01:37:37.040451 kernel: pci 0000:04:00.0: Adding to iommu group 16 Nov 8 01:37:37.040500 kernel: pci 0000:06:00.0: Adding to iommu group 17 Nov 8 01:37:37.040554 kernel: pci 0000:07:00.0: Adding to iommu group 17 Nov 8 01:37:37.040562 kernel: DMAR: Intel(R) Virtualization Technology for Directed I/O Nov 8 01:37:37.040568 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Nov 8 01:37:37.040576 kernel: software IO TLB: mapped [mem 0x0000000086fce000-0x000000008afce000] (64MB) Nov 8 01:37:37.040582 kernel: RAPL PMU: API unit is 2^-32 Joules, 3 fixed counters, 655360 ms ovfl timer Nov 8 01:37:37.040588 kernel: RAPL PMU: hw unit of domain pp0-core 2^-14 Joules Nov 8 01:37:37.040594 kernel: RAPL PMU: hw unit of domain package 2^-14 Joules Nov 8 01:37:37.040599 kernel: RAPL PMU: hw unit of domain dram 2^-14 Joules Nov 8 01:37:37.040651 kernel: platform rtc_cmos: registered platform RTC device (no PNP device found) Nov 8 01:37:37.040660 kernel: Initialise system trusted keyrings Nov 8 01:37:37.040666 kernel: workingset: timestamp_bits=39 max_order=23 bucket_order=0 Nov 8 01:37:37.040673 kernel: Key type asymmetric registered Nov 8 01:37:37.040679 kernel: Asymmetric key parser 'x509' registered Nov 8 01:37:37.040685 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Nov 8 01:37:37.040690 kernel: io scheduler mq-deadline registered Nov 8 01:37:37.040696 kernel: io scheduler kyber registered Nov 8 01:37:37.040702 kernel: io scheduler bfq registered Nov 8 01:37:37.040750 kernel: pcieport 0000:00:01.0: PME: Signaling with IRQ 121 Nov 8 01:37:37.040801 kernel: pcieport 0000:00:1b.0: PME: Signaling with IRQ 122 Nov 8 01:37:37.040850 kernel: pcieport 0000:00:1b.4: PME: Signaling with IRQ 123 Nov 8 01:37:37.040901 kernel: pcieport 0000:00:1b.5: PME: Signaling with IRQ 124 Nov 8 01:37:37.040949 kernel: pcieport 0000:00:1c.0: PME: Signaling with IRQ 125 Nov 8 01:37:37.040998 kernel: pcieport 0000:00:1c.3: PME: Signaling with IRQ 126 Nov 8 01:37:37.041052 kernel: thermal LNXTHERM:00: registered as thermal_zone0 Nov 8 01:37:37.041060 kernel: ACPI: thermal: Thermal Zone [TZ00] (28 C) Nov 8 01:37:37.041066 kernel: ERST: Error Record Serialization Table (ERST) support is initialized. Nov 8 01:37:37.041072 kernel: pstore: Using crash dump compression: deflate Nov 8 01:37:37.041080 kernel: pstore: Registered erst as persistent store backend Nov 8 01:37:37.041085 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Nov 8 01:37:37.041091 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 8 01:37:37.041097 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Nov 8 01:37:37.041103 kernel: 00:03: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Nov 8 01:37:37.041109 kernel: hpet_acpi_add: no address or irqs in _CRS Nov 8 01:37:37.041188 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x1B, rev-id 16) Nov 8 01:37:37.041197 kernel: i8042: PNP: No PS/2 controller found. Nov 8 01:37:37.041262 kernel: rtc_cmos rtc_cmos: RTC can wake from S4 Nov 8 01:37:37.041308 kernel: rtc_cmos rtc_cmos: registered as rtc0 Nov 8 01:37:37.041353 kernel: rtc_cmos rtc_cmos: setting system clock to 2025-11-08T01:37:35 UTC (1762565855) Nov 8 01:37:37.041398 kernel: rtc_cmos rtc_cmos: alarms up to one month, y3k, 114 bytes nvram Nov 8 01:37:37.041406 kernel: intel_pstate: Intel P-state driver initializing Nov 8 01:37:37.041412 kernel: intel_pstate: Disabling energy efficiency optimization Nov 8 01:37:37.041418 kernel: intel_pstate: HWP enabled Nov 8 01:37:37.041424 kernel: vesafb: mode is 1024x768x8, linelength=1024, pages=0 Nov 8 01:37:37.041430 kernel: vesafb: scrolling: redraw Nov 8 01:37:37.041437 kernel: vesafb: Pseudocolor: size=0:8:8:8, shift=0:0:0:0 Nov 8 01:37:37.041443 kernel: vesafb: framebuffer at 0x94000000, mapped to 0x000000009517b72d, using 768k, total 768k Nov 8 01:37:37.041449 kernel: Console: switching to colour frame buffer device 128x48 Nov 8 01:37:37.041454 kernel: fb0: VESA VGA frame buffer device Nov 8 01:37:37.041460 kernel: NET: Registered PF_INET6 protocol family Nov 8 01:37:37.041466 kernel: Segment Routing with IPv6 Nov 8 01:37:37.041472 kernel: In-situ OAM (IOAM) with IPv6 Nov 8 01:37:37.041478 kernel: NET: Registered PF_PACKET protocol family Nov 8 01:37:37.041483 kernel: Key type dns_resolver registered Nov 8 01:37:37.041490 kernel: microcode: Current revision: 0x00000102 Nov 8 01:37:37.041496 kernel: microcode: Microcode Update Driver: v2.2. Nov 8 01:37:37.041501 kernel: IPI shorthand broadcast: enabled Nov 8 01:37:37.041507 kernel: sched_clock: Marking stable (1561085242, 1369070988)->(4401410064, -1471253834) Nov 8 01:37:37.041513 kernel: registered taskstats version 1 Nov 8 01:37:37.041519 kernel: Loading compiled-in X.509 certificates Nov 8 01:37:37.041525 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.113-flatcar: cf7a35a152685ec84a621291e4ce58c959319dfd' Nov 8 01:37:37.041530 kernel: Key type .fscrypt registered Nov 8 01:37:37.041536 kernel: Key type fscrypt-provisioning registered Nov 8 01:37:37.041542 kernel: ima: Allocated hash algorithm: sha1 Nov 8 01:37:37.041548 kernel: ima: No architecture policies found Nov 8 01:37:37.041554 kernel: clk: Disabling unused clocks Nov 8 01:37:37.041560 kernel: Freeing unused kernel image (initmem) memory: 42880K Nov 8 01:37:37.041565 kernel: Write protecting the kernel read-only data: 36864k Nov 8 01:37:37.041571 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Nov 8 01:37:37.041577 kernel: Run /init as init process Nov 8 01:37:37.041583 kernel: with arguments: Nov 8 01:37:37.041589 kernel: /init Nov 8 01:37:37.041595 kernel: with environment: Nov 8 01:37:37.041601 kernel: HOME=/ Nov 8 01:37:37.041606 kernel: TERM=linux Nov 8 01:37:37.041613 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Nov 8 01:37:37.041621 systemd[1]: Detected architecture x86-64. Nov 8 01:37:37.041627 systemd[1]: Running in initrd. Nov 8 01:37:37.041632 systemd[1]: No hostname configured, using default hostname. Nov 8 01:37:37.041639 systemd[1]: Hostname set to . Nov 8 01:37:37.041645 systemd[1]: Initializing machine ID from random generator. Nov 8 01:37:37.041651 systemd[1]: Queued start job for default target initrd.target. Nov 8 01:37:37.041657 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 8 01:37:37.041663 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 8 01:37:37.041669 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Nov 8 01:37:37.041675 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 8 01:37:37.041681 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Nov 8 01:37:37.041688 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Nov 8 01:37:37.041695 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Nov 8 01:37:37.041701 kernel: tsc: Refined TSC clocksource calibration: 3407.998 MHz Nov 8 01:37:37.041707 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Nov 8 01:37:37.041713 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd208cfc, max_idle_ns: 440795283699 ns Nov 8 01:37:37.041719 kernel: clocksource: Switched to clocksource tsc Nov 8 01:37:37.041725 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 8 01:37:37.041732 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 8 01:37:37.041738 systemd[1]: Reached target paths.target - Path Units. Nov 8 01:37:37.041744 systemd[1]: Reached target slices.target - Slice Units. Nov 8 01:37:37.041750 systemd[1]: Reached target swap.target - Swaps. Nov 8 01:37:37.041756 systemd[1]: Reached target timers.target - Timer Units. Nov 8 01:37:37.041762 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Nov 8 01:37:37.041768 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 8 01:37:37.041774 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 8 01:37:37.041780 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Nov 8 01:37:37.041787 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 8 01:37:37.041793 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 8 01:37:37.041799 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 8 01:37:37.041805 systemd[1]: Reached target sockets.target - Socket Units. Nov 8 01:37:37.041811 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Nov 8 01:37:37.041817 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 8 01:37:37.041823 systemd[1]: Finished network-cleanup.service - Network Cleanup. Nov 8 01:37:37.041829 systemd[1]: Starting systemd-fsck-usr.service... Nov 8 01:37:37.041836 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 8 01:37:37.041852 systemd-journald[266]: Collecting audit messages is disabled. Nov 8 01:37:37.041866 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 8 01:37:37.041873 systemd-journald[266]: Journal started Nov 8 01:37:37.041887 systemd-journald[266]: Runtime Journal (/run/log/journal/58bc4733a404490d8afa8ebd88ef6673) is 8.0M, max 639.9M, 631.9M free. Nov 8 01:37:37.055933 systemd-modules-load[267]: Inserted module 'overlay' Nov 8 01:37:37.084246 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 8 01:37:37.132190 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 8 01:37:37.132221 systemd[1]: Started systemd-journald.service - Journal Service. Nov 8 01:37:37.150936 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Nov 8 01:37:37.151028 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 8 01:37:37.151118 systemd[1]: Finished systemd-fsck-usr.service. Nov 8 01:37:37.152231 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 8 01:37:37.169514 systemd-modules-load[267]: Inserted module 'br_netfilter' Nov 8 01:37:37.169924 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 8 01:37:37.258514 kernel: Bridge firewalling registered Nov 8 01:37:37.237732 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 8 01:37:37.259034 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:37:37.290695 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 8 01:37:37.308560 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 8 01:37:37.347466 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 8 01:37:37.348382 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 8 01:37:37.349242 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 8 01:37:37.357312 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 8 01:37:37.357665 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 8 01:37:37.358645 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 8 01:37:37.361373 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:37:37.361943 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Nov 8 01:37:37.376308 systemd-resolved[304]: Positive Trust Anchors: Nov 8 01:37:37.376313 systemd-resolved[304]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 8 01:37:37.376339 systemd-resolved[304]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 8 01:37:37.378002 systemd-resolved[304]: Defaulting to hostname 'linux'. Nov 8 01:37:37.394452 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 8 01:37:37.419443 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 8 01:37:37.524263 dracut-cmdline[306]: dracut-dracut-053 Nov 8 01:37:37.524263 dracut-cmdline[306]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected flatcar.oem.id=packet flatcar.autologin verity.usrhash=480a02cf7f2001774aa495c3e719d4173e968e6839485a7d2b207ef2facca472 Nov 8 01:37:37.586201 kernel: SCSI subsystem initialized Nov 8 01:37:37.609178 kernel: Loading iSCSI transport class v2.0-870. Nov 8 01:37:37.632185 kernel: iscsi: registered transport (tcp) Nov 8 01:37:37.665381 kernel: iscsi: registered transport (qla4xxx) Nov 8 01:37:37.665398 kernel: QLogic iSCSI HBA Driver Nov 8 01:37:37.698040 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Nov 8 01:37:37.725410 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Nov 8 01:37:37.781461 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 8 01:37:37.781483 kernel: device-mapper: uevent: version 1.0.3 Nov 8 01:37:37.801219 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Nov 8 01:37:37.859193 kernel: raid6: avx2x4 gen() 53148 MB/s Nov 8 01:37:37.891208 kernel: raid6: avx2x2 gen() 53274 MB/s Nov 8 01:37:37.927589 kernel: raid6: avx2x1 gen() 45177 MB/s Nov 8 01:37:37.927605 kernel: raid6: using algorithm avx2x2 gen() 53274 MB/s Nov 8 01:37:37.975671 kernel: raid6: .... xor() 31128 MB/s, rmw enabled Nov 8 01:37:37.975688 kernel: raid6: using avx2x2 recovery algorithm Nov 8 01:37:38.017183 kernel: xor: automatically using best checksumming function avx Nov 8 01:37:38.135150 kernel: Btrfs loaded, zoned=no, fsverity=no Nov 8 01:37:38.140484 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Nov 8 01:37:38.173538 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 8 01:37:38.180554 systemd-udevd[492]: Using default interface naming scheme 'v255'. Nov 8 01:37:38.184228 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 8 01:37:38.228408 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Nov 8 01:37:38.265848 dracut-pre-trigger[504]: rd.md=0: removing MD RAID activation Nov 8 01:37:38.282972 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Nov 8 01:37:38.313520 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 8 01:37:38.414997 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 8 01:37:38.449206 kernel: pps_core: LinuxPPS API ver. 1 registered Nov 8 01:37:38.449264 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Nov 8 01:37:38.477184 kernel: cryptd: max_cpu_qlen set to 1000 Nov 8 01:37:38.500821 kernel: ACPI: bus type USB registered Nov 8 01:37:38.500840 kernel: usbcore: registered new interface driver usbfs Nov 8 01:37:38.516025 kernel: usbcore: registered new interface driver hub Nov 8 01:37:38.530815 kernel: usbcore: registered new device driver usb Nov 8 01:37:38.546145 kernel: PTP clock support registered Nov 8 01:37:38.546163 kernel: libata version 3.00 loaded. Nov 8 01:37:38.562812 kernel: ahci 0000:00:17.0: version 3.0 Nov 8 01:37:38.562985 kernel: AVX2 version of gcm_enc/dec engaged. Nov 8 01:37:38.564466 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Nov 8 01:37:38.641566 kernel: ahci 0000:00:17.0: AHCI 0001.0301 32 slots 7 ports 6 Gbps 0x7f impl SATA mode Nov 8 01:37:38.641786 kernel: ahci 0000:00:17.0: flags: 64bit ncq sntf clo only pio slum part ems deso sadm sds apst Nov 8 01:37:38.641954 kernel: scsi host0: ahci Nov 8 01:37:38.642132 kernel: scsi host1: ahci Nov 8 01:37:38.642326 kernel: scsi host2: ahci Nov 8 01:37:38.627515 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Nov 8 01:37:38.669246 kernel: scsi host3: ahci Nov 8 01:37:38.669526 kernel: scsi host4: ahci Nov 8 01:37:38.656853 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Nov 8 01:37:38.715796 kernel: scsi host5: ahci Nov 8 01:37:38.716140 kernel: scsi host6: ahci Nov 8 01:37:38.716451 kernel: ata1: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516100 irq 127 Nov 8 01:37:38.685827 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 8 01:37:38.850246 kernel: ata2: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516180 irq 127 Nov 8 01:37:38.850260 kernel: ata3: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516200 irq 127 Nov 8 01:37:38.850268 kernel: ata4: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516280 irq 127 Nov 8 01:37:38.850276 kernel: ata5: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516300 irq 127 Nov 8 01:37:38.850283 kernel: ata6: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516380 irq 127 Nov 8 01:37:38.850290 kernel: ata7: SATA max UDMA/133 abar m2048@0x95516000 port 0x95516400 irq 127 Nov 8 01:37:38.850298 kernel: AES CTR mode by8 optimization enabled Nov 8 01:37:38.850305 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Nov 8 01:37:38.836242 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 8 01:37:38.875312 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Nov 8 01:37:38.875247 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 8 01:37:38.953190 kernel: igb 0000:03:00.0: added PHC on eth0 Nov 8 01:37:38.953280 kernel: igb 0000:03:00.0: Intel(R) Gigabit Ethernet Network Connection Nov 8 01:37:38.953357 kernel: igb 0000:03:00.0: eth0: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:70:d3:7e Nov 8 01:37:38.953430 kernel: igb 0000:03:00.0: eth0: PBA No: 010000-000 Nov 8 01:37:38.953500 kernel: igb 0000:03:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Nov 8 01:37:38.875280 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:37:39.068988 kernel: mlx5_core 0000:01:00.0: firmware version: 14.31.1014 Nov 8 01:37:39.069080 kernel: mlx5_core 0000:01:00.0: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Nov 8 01:37:39.069159 kernel: igb 0000:04:00.0: added PHC on eth1 Nov 8 01:37:39.069231 kernel: igb 0000:04:00.0: Intel(R) Gigabit Ethernet Network Connection Nov 8 01:37:39.069297 kernel: igb 0000:04:00.0: eth1: (PCIe:2.5Gb/s:Width x1) 3c:ec:ef:70:d3:7f Nov 8 01:37:39.069362 kernel: igb 0000:04:00.0: eth1: PBA No: 010000-000 Nov 8 01:37:39.069427 kernel: ata3: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.069436 kernel: igb 0000:04:00.0: Using MSI-X interrupts. 4 rx queue(s), 4 tx queue(s) Nov 8 01:37:39.069499 kernel: ata5: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:38.969257 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 8 01:37:39.096599 kernel: ata2: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Nov 8 01:37:39.096612 kernel: ata4: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.111144 kernel: ata1: SATA link up 6.0 Gbps (SStatus 133 SControl 300) Nov 8 01:37:39.126145 kernel: ata6: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.139177 kernel: ata7: SATA link down (SStatus 0 SControl 300) Nov 8 01:37:39.152197 kernel: ata2.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Nov 8 01:37:39.166181 kernel: ata1.00: ATA-11: Micron_5300_MTFDDAK480TDT, D3MU001, max UDMA/133 Nov 8 01:37:39.182274 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Nov 8 01:37:39.235500 kernel: ata2.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Nov 8 01:37:39.235512 kernel: ata1.00: 937703088 sectors, multi 16: LBA48 NCQ (depth 32), AA Nov 8 01:37:39.235520 kernel: ata1.00: Features: NCQ-prio Nov 8 01:37:39.235528 kernel: ata2.00: Features: NCQ-prio Nov 8 01:37:39.198239 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 8 01:37:39.347213 kernel: ata1.00: configured for UDMA/133 Nov 8 01:37:39.347232 kernel: ata2.00: configured for UDMA/133 Nov 8 01:37:39.347242 kernel: scsi 0:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Nov 8 01:37:39.347348 kernel: mlx5_core 0000:01:00.0: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Nov 8 01:37:39.347438 kernel: mlx5_core 0000:01:00.0: Port module event: module 0, Cable plugged Nov 8 01:37:39.347516 kernel: scsi 1:0:0:0: Direct-Access ATA Micron_5300_MTFD U001 PQ: 0 ANSI: 5 Nov 8 01:37:39.347595 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Nov 8 01:37:39.347677 kernel: igb 0000:04:00.0 eno2: renamed from eth1 Nov 8 01:37:39.347755 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 1 Nov 8 01:37:39.198269 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:37:39.252741 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Nov 8 01:37:39.376146 kernel: xhci_hcd 0000:00:14.0: hcc params 0x200077c1 hci version 0x110 quirks 0x0000000000009810 Nov 8 01:37:39.376286 kernel: igb 0000:03:00.0 eno1: renamed from eth0 Nov 8 01:37:39.381201 kernel: xhci_hcd 0000:00:14.0: xHCI Host Controller Nov 8 01:37:39.418234 kernel: xhci_hcd 0000:00:14.0: new USB bus registered, assigned bus number 2 Nov 8 01:37:39.427183 kernel: xhci_hcd 0000:00:14.0: Host supports USB 3.1 Enhanced SuperSpeed Nov 8 01:37:39.427284 kernel: hub 1-0:1.0: USB hub found Nov 8 01:37:39.459060 kernel: hub 1-0:1.0: 16 ports detected Nov 8 01:37:39.472226 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 8 01:37:39.513641 kernel: hub 2-0:1.0: USB hub found Nov 8 01:37:39.513788 kernel: hub 2-0:1.0: 10 ports detected Nov 8 01:37:39.513867 kernel: mlx5_core 0000:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Nov 8 01:37:39.494424 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Nov 8 01:37:39.764515 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:39.764529 kernel: mlx5_core 0000:01:00.1: firmware version: 14.31.1014 Nov 8 01:37:39.764619 kernel: ata2.00: Enabling discard_zeroes_data Nov 8 01:37:39.764628 kernel: sd 0:0:0:0: [sda] 937703088 512-byte logical blocks: (480 GB/447 GiB) Nov 8 01:37:39.764703 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Nov 8 01:37:39.764767 kernel: mlx5_core 0000:01:00.1: 63.008 Gb/s available PCIe bandwidth (8.0 GT/s PCIe x8 link) Nov 8 01:37:39.764835 kernel: sd 0:0:0:0: [sda] Write Protect is off Nov 8 01:37:39.764902 kernel: sd 1:0:0:0: [sdb] 937703088 512-byte logical blocks: (480 GB/447 GiB) Nov 8 01:37:39.764969 kernel: sd 1:0:0:0: [sdb] 4096-byte physical blocks Nov 8 01:37:39.765031 kernel: sd 1:0:0:0: [sdb] Write Protect is off Nov 8 01:37:39.765093 kernel: sd 1:0:0:0: [sdb] Mode Sense: 00 3a 00 00 Nov 8 01:37:39.765164 kernel: sd 1:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Nov 8 01:37:39.765257 kernel: sd 1:0:0:0: [sdb] Preferred minimum I/O size 4096 bytes Nov 8 01:37:39.765331 kernel: ata2.00: Enabling discard_zeroes_data Nov 8 01:37:39.765343 kernel: sd 1:0:0:0: [sdb] Attached SCSI disk Nov 8 01:37:39.765410 kernel: usb 1-14: new high-speed USB device number 2 using xhci_hcd Nov 8 01:37:39.765427 kernel: sd 0:0:0:0: [sda] Mode Sense: 00 3a 00 00 Nov 8 01:37:39.765489 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Nov 8 01:37:39.765552 kernel: sd 0:0:0:0: [sda] Preferred minimum I/O size 4096 bytes Nov 8 01:37:39.796951 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:39.821594 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 8 01:37:39.821611 kernel: GPT:9289727 != 937703087 Nov 8 01:37:39.837602 kernel: GPT:Alternate GPT header not at the end of the disk. Nov 8 01:37:39.837618 kernel: mlx5_core 0000:01:00.1: E-Switch: Total vports 10, per vport: max uc(128) max mc(2048) Nov 8 01:37:39.837702 kernel: GPT:9289727 != 937703087 Nov 8 01:37:39.837711 kernel: GPT: Use GNU Parted to correct GPT errors. Nov 8 01:37:39.837718 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:39.855147 kernel: mlx5_core 0000:01:00.1: Port module event: module 1, Cable plugged Nov 8 01:37:39.855238 kernel: hub 1-14:1.0: USB hub found Nov 8 01:37:39.855323 kernel: hub 1-14:1.0: 4 ports detected Nov 8 01:37:39.864794 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Nov 8 01:37:39.970174 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:37:40.006853 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Micron_5300_MTFDDAK480TDT EFI-SYSTEM. Nov 8 01:37:40.071401 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by (udev-worker) (669) Nov 8 01:37:40.071417 kernel: BTRFS: device fsid a2737782-a37e-42f9-8b56-489a87f47acc devid 1 transid 35 /dev/sda3 scanned by (udev-worker) (663) Nov 8 01:37:40.061950 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Micron_5300_MTFDDAK480TDT ROOT. Nov 8 01:37:40.085985 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Nov 8 01:37:40.118211 kernel: mlx5_core 0000:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Nov 8 01:37:40.134191 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Micron_5300_MTFDDAK480TDT USR-A. Nov 8 01:37:40.203413 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: renamed from eth1 Nov 8 01:37:40.203508 kernel: usb 1-14.1: new low-speed USB device number 3 using xhci_hcd Nov 8 01:37:40.203526 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: renamed from eth0 Nov 8 01:37:40.156264 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Micron_5300_MTFDDAK480TDT USR-A. Nov 8 01:37:40.205451 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Nov 8 01:37:40.261194 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:40.261208 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:40.261219 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:40.237672 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 8 01:37:40.262216 disk-uuid[723]: Primary Header is updated. Nov 8 01:37:40.262216 disk-uuid[723]: Secondary Entries is updated. Nov 8 01:37:40.262216 disk-uuid[723]: Secondary Header is updated. Nov 8 01:37:40.315596 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:40.315607 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:40.315614 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:40.340149 kernel: hid: raw HID events driver (C) Jiri Kosina Nov 8 01:37:40.363274 kernel: usbcore: registered new interface driver usbhid Nov 8 01:37:40.363348 kernel: usbhid: USB HID core driver Nov 8 01:37:40.410286 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.0/0003:0557:2419.0001/input/input0 Nov 8 01:37:40.423624 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:37:40.508665 kernel: hid-generic 0003:0557:2419.0001: input,hidraw0: USB HID v1.00 Keyboard [HID 0557:2419] on usb-0000:00:14.0-14.1/input0 Nov 8 01:37:40.508790 kernel: input: HID 0557:2419 as /devices/pci0000:00/0000:00:14.0/usb1/1-14/1-14.1/1-14.1:1.1/0003:0557:2419.0002/input/input1 Nov 8 01:37:40.544027 kernel: hid-generic 0003:0557:2419.0002: input,hidraw1: USB HID v1.00 Mouse [HID 0557:2419] on usb-0000:00:14.0-14.1/input1 Nov 8 01:37:41.297131 kernel: ata1.00: Enabling discard_zeroes_data Nov 8 01:37:41.317129 disk-uuid[724]: The operation has completed successfully. Nov 8 01:37:41.325257 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Nov 8 01:37:41.354135 systemd[1]: disk-uuid.service: Deactivated successfully. Nov 8 01:37:41.354192 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Nov 8 01:37:41.393386 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Nov 8 01:37:41.418325 sh[751]: Success Nov 8 01:37:41.428245 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Nov 8 01:37:41.474509 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Nov 8 01:37:41.495293 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Nov 8 01:37:41.503544 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Nov 8 01:37:41.565055 kernel: BTRFS info (device dm-0): first mount of filesystem a2737782-a37e-42f9-8b56-489a87f47acc Nov 8 01:37:41.565075 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:41.587034 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Nov 8 01:37:41.606571 kernel: BTRFS info (device dm-0): disabling log replay at mount time Nov 8 01:37:41.624983 kernel: BTRFS info (device dm-0): using free space tree Nov 8 01:37:41.663190 kernel: BTRFS info (device dm-0): enabling ssd optimizations Nov 8 01:37:41.665003 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Nov 8 01:37:41.673594 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Nov 8 01:37:41.688314 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Nov 8 01:37:41.700967 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Nov 8 01:37:41.812605 kernel: BTRFS info (device sda6): first mount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:41.812631 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:41.812646 kernel: BTRFS info (device sda6): using free space tree Nov 8 01:37:41.812654 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 8 01:37:41.812661 kernel: BTRFS info (device sda6): auto enabling async discard Nov 8 01:37:41.836223 kernel: BTRFS info (device sda6): last unmount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:41.845330 systemd[1]: Finished ignition-setup.service - Ignition (setup). Nov 8 01:37:41.874329 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Nov 8 01:37:41.891801 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 8 01:37:41.920271 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 8 01:37:41.931824 systemd-networkd[934]: lo: Link UP Nov 8 01:37:41.931827 systemd-networkd[934]: lo: Gained carrier Nov 8 01:37:41.940648 ignition[859]: Ignition 2.19.0 Nov 8 01:37:41.935073 systemd-networkd[934]: Enumeration completed Nov 8 01:37:41.940653 ignition[859]: Stage: fetch-offline Nov 8 01:37:41.935149 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 8 01:37:41.940675 ignition[859]: no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:41.936000 systemd-networkd[934]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:41.940680 ignition[859]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:41.942929 unknown[859]: fetched base config from "system" Nov 8 01:37:41.940733 ignition[859]: parsed url from cmdline: "" Nov 8 01:37:41.942933 unknown[859]: fetched user config from "system" Nov 8 01:37:41.940735 ignition[859]: no config URL provided Nov 8 01:37:41.947571 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Nov 8 01:37:41.940737 ignition[859]: reading system config file "/usr/lib/ignition/user.ign" Nov 8 01:37:41.964887 systemd-networkd[934]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:41.940761 ignition[859]: parsing config with SHA512: 73ad4a693266b007874f0d28e1091c13061053cdd7b48746fdbadf195e470ff304d4cbb2e51232992df41ca9e71db2bb394fceb2ae45c785dd19a5ae480522f7 Nov 8 01:37:41.966649 systemd[1]: Reached target network.target - Network. Nov 8 01:37:41.943204 ignition[859]: fetch-offline: fetch-offline passed Nov 8 01:37:41.981323 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Nov 8 01:37:41.943206 ignition[859]: POST message to Packet Timeline Nov 8 01:37:41.990324 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Nov 8 01:37:41.943209 ignition[859]: POST Status error: resource requires networking Nov 8 01:37:41.993436 systemd-networkd[934]: enp1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:41.943259 ignition[859]: Ignition finished successfully Nov 8 01:37:42.001986 ignition[947]: Ignition 2.19.0 Nov 8 01:37:42.001994 ignition[947]: Stage: kargs Nov 8 01:37:42.002201 ignition[947]: no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:42.002215 ignition[947]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:42.003345 ignition[947]: kargs: kargs passed Nov 8 01:37:42.227291 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Nov 8 01:37:42.221665 systemd-networkd[934]: enp1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 8 01:37:42.003350 ignition[947]: POST message to Packet Timeline Nov 8 01:37:42.003366 ignition[947]: GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:37:42.004237 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:41411->[::1]:53: read: connection refused Nov 8 01:37:42.204476 ignition[947]: GET https://metadata.packet.net/metadata: attempt #2 Nov 8 01:37:42.204969 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:55986->[::1]:53: read: connection refused Nov 8 01:37:42.510193 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Nov 8 01:37:42.510825 systemd-networkd[934]: eno1: Link UP Nov 8 01:37:42.511046 systemd-networkd[934]: eno2: Link UP Nov 8 01:37:42.511222 systemd-networkd[934]: enp1s0f0np0: Link UP Nov 8 01:37:42.511413 systemd-networkd[934]: enp1s0f0np0: Gained carrier Nov 8 01:37:42.520425 systemd-networkd[934]: enp1s0f1np1: Link UP Nov 8 01:37:42.549313 systemd-networkd[934]: enp1s0f0np0: DHCPv4 address 139.178.94.189/31, gateway 139.178.94.188 acquired from 145.40.83.140 Nov 8 01:37:42.605240 ignition[947]: GET https://metadata.packet.net/metadata: attempt #3 Nov 8 01:37:42.606374 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:52706->[::1]:53: read: connection refused Nov 8 01:37:43.265883 systemd-networkd[934]: enp1s0f1np1: Gained carrier Nov 8 01:37:43.407567 ignition[947]: GET https://metadata.packet.net/metadata: attempt #4 Nov 8 01:37:43.408720 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:42734->[::1]:53: read: connection refused Nov 8 01:37:43.649767 systemd-networkd[934]: enp1s0f0np0: Gained IPv6LL Nov 8 01:37:45.010493 ignition[947]: GET https://metadata.packet.net/metadata: attempt #5 Nov 8 01:37:45.011603 ignition[947]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:46313->[::1]:53: read: connection refused Nov 8 01:37:45.121758 systemd-networkd[934]: enp1s0f1np1: Gained IPv6LL Nov 8 01:37:48.215234 ignition[947]: GET https://metadata.packet.net/metadata: attempt #6 Nov 8 01:37:52.062805 ignition[947]: GET result: OK Nov 8 01:37:53.303309 ignition[947]: Ignition finished successfully Nov 8 01:37:53.308523 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Nov 8 01:37:53.337391 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Nov 8 01:37:53.343585 ignition[966]: Ignition 2.19.0 Nov 8 01:37:53.343589 ignition[966]: Stage: disks Nov 8 01:37:53.343702 ignition[966]: no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:53.343709 ignition[966]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:53.344284 ignition[966]: disks: disks passed Nov 8 01:37:53.344286 ignition[966]: POST message to Packet Timeline Nov 8 01:37:53.344296 ignition[966]: GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:37:54.760482 ignition[966]: GET result: OK Nov 8 01:37:55.287621 ignition[966]: Ignition finished successfully Nov 8 01:37:55.290803 systemd[1]: Finished ignition-disks.service - Ignition (disks). Nov 8 01:37:55.306407 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Nov 8 01:37:55.324410 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 8 01:37:55.345405 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 8 01:37:55.366564 systemd[1]: Reached target sysinit.target - System Initialization. Nov 8 01:37:55.386440 systemd[1]: Reached target basic.target - Basic System. Nov 8 01:37:55.415418 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Nov 8 01:37:55.455093 systemd-fsck[984]: ROOT: clean, 14/553520 files, 52654/553472 blocks Nov 8 01:37:55.465598 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Nov 8 01:37:55.501409 systemd[1]: Mounting sysroot.mount - /sysroot... Nov 8 01:37:55.603152 kernel: EXT4-fs (sda9): mounted filesystem 3cd35b5c-4e0e-45c1-abc9-cf70eebd42df r/w with ordered data mode. Quota mode: none. Nov 8 01:37:55.603666 systemd[1]: Mounted sysroot.mount - /sysroot. Nov 8 01:37:55.613571 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Nov 8 01:37:55.639354 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 8 01:37:55.639927 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Nov 8 01:37:55.662921 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Nov 8 01:37:55.685146 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (993) Nov 8 01:37:55.709164 kernel: BTRFS info (device sda6): first mount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:55.709200 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:55.709208 kernel: BTRFS info (device sda6): using free space tree Nov 8 01:37:55.734696 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Nov 8 01:37:55.798447 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 8 01:37:55.798462 kernel: BTRFS info (device sda6): auto enabling async discard Nov 8 01:37:55.798470 coreos-metadata[995]: Nov 08 01:37:55.738 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Nov 8 01:37:55.798378 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Nov 8 01:37:55.798396 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Nov 8 01:37:55.869353 coreos-metadata[1013]: Nov 08 01:37:55.854 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Nov 8 01:37:55.819509 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 8 01:37:55.858412 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Nov 8 01:37:55.894411 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Nov 8 01:37:55.925514 initrd-setup-root[1025]: cut: /sysroot/etc/passwd: No such file or directory Nov 8 01:37:55.935257 initrd-setup-root[1032]: cut: /sysroot/etc/group: No such file or directory Nov 8 01:37:55.945478 initrd-setup-root[1039]: cut: /sysroot/etc/shadow: No such file or directory Nov 8 01:37:55.955206 initrd-setup-root[1046]: cut: /sysroot/etc/gshadow: No such file or directory Nov 8 01:37:55.991664 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Nov 8 01:37:56.015417 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Nov 8 01:37:56.051360 kernel: BTRFS info (device sda6): last unmount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:56.033783 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Nov 8 01:37:56.059918 systemd[1]: sysroot-oem.mount: Deactivated successfully. Nov 8 01:37:56.081380 ignition[1114]: INFO : Ignition 2.19.0 Nov 8 01:37:56.081380 ignition[1114]: INFO : Stage: mount Nov 8 01:37:56.095377 ignition[1114]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:56.095377 ignition[1114]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:56.095377 ignition[1114]: INFO : mount: mount passed Nov 8 01:37:56.095377 ignition[1114]: INFO : POST message to Packet Timeline Nov 8 01:37:56.095377 ignition[1114]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:37:56.089479 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Nov 8 01:37:56.654305 coreos-metadata[995]: Nov 08 01:37:56.654 INFO Fetch successful Nov 8 01:37:56.734943 coreos-metadata[995]: Nov 08 01:37:56.734 INFO wrote hostname ci-4081.3.6-n-3863d1ecf5 to /sysroot/etc/hostname Nov 8 01:37:56.736459 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Nov 8 01:37:56.803652 coreos-metadata[1013]: Nov 08 01:37:56.803 INFO Fetch successful Nov 8 01:37:56.884580 systemd[1]: flatcar-static-network.service: Deactivated successfully. Nov 8 01:37:56.884639 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Nov 8 01:37:57.156290 ignition[1114]: INFO : GET result: OK Nov 8 01:37:58.389560 ignition[1114]: INFO : Ignition finished successfully Nov 8 01:37:58.393418 systemd[1]: Finished ignition-mount.service - Ignition (mount). Nov 8 01:37:58.423359 systemd[1]: Starting ignition-files.service - Ignition (files)... Nov 8 01:37:58.433372 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 8 01:37:58.488200 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (1137) Nov 8 01:37:58.518038 kernel: BTRFS info (device sda6): first mount of filesystem 7b59d8a2-cf4e-4d67-8d1e-00d7f134f45e Nov 8 01:37:58.518054 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Nov 8 01:37:58.536209 kernel: BTRFS info (device sda6): using free space tree Nov 8 01:37:58.575522 kernel: BTRFS info (device sda6): enabling ssd optimizations Nov 8 01:37:58.575538 kernel: BTRFS info (device sda6): auto enabling async discard Nov 8 01:37:58.589523 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 8 01:37:58.618642 ignition[1154]: INFO : Ignition 2.19.0 Nov 8 01:37:58.618642 ignition[1154]: INFO : Stage: files Nov 8 01:37:58.633385 ignition[1154]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 8 01:37:58.633385 ignition[1154]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:37:58.633385 ignition[1154]: DEBUG : files: compiled without relabeling support, skipping Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Nov 8 01:37:58.633385 ignition[1154]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 8 01:37:58.633385 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Nov 8 01:37:58.622968 unknown[1154]: wrote ssh authorized keys file for user: core Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:58.797413 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:59.046404 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Nov 8 01:37:59.181257 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Nov 8 01:37:59.574837 ignition[1154]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 8 01:37:59.574837 ignition[1154]: INFO : files: op(c): [started] processing unit "containerd.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(c): [finished] processing unit "containerd.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Nov 8 01:37:59.605414 ignition[1154]: INFO : files: files passed Nov 8 01:37:59.605414 ignition[1154]: INFO : POST message to Packet Timeline Nov 8 01:37:59.605414 ignition[1154]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:38:00.621738 ignition[1154]: INFO : GET result: OK Nov 8 01:38:01.755074 ignition[1154]: INFO : Ignition finished successfully Nov 8 01:38:01.758235 systemd[1]: Finished ignition-files.service - Ignition (files). Nov 8 01:38:01.797425 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Nov 8 01:38:01.807764 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Nov 8 01:38:01.828696 systemd[1]: ignition-quench.service: Deactivated successfully. Nov 8 01:38:01.828787 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Nov 8 01:38:01.868103 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 8 01:38:01.880654 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Nov 8 01:38:01.912370 initrd-setup-root-after-ignition[1193]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 8 01:38:01.912370 initrd-setup-root-after-ignition[1193]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Nov 8 01:38:01.926360 initrd-setup-root-after-ignition[1197]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 8 01:38:01.919359 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Nov 8 01:38:01.978827 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 8 01:38:01.978883 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Nov 8 01:38:01.997642 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Nov 8 01:38:02.018314 systemd[1]: Reached target initrd.target - Initrd Default Target. Nov 8 01:38:02.042612 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Nov 8 01:38:02.053519 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Nov 8 01:38:02.119772 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 8 01:38:02.145566 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Nov 8 01:38:02.209436 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Nov 8 01:38:02.221772 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 8 01:38:02.242844 systemd[1]: Stopped target timers.target - Timer Units. Nov 8 01:38:02.263876 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 8 01:38:02.264313 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 8 01:38:02.292983 systemd[1]: Stopped target initrd.target - Initrd Default Target. Nov 8 01:38:02.313783 systemd[1]: Stopped target basic.target - Basic System. Nov 8 01:38:02.334869 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Nov 8 01:38:02.355867 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Nov 8 01:38:02.376759 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Nov 8 01:38:02.397777 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Nov 8 01:38:02.417876 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Nov 8 01:38:02.438796 systemd[1]: Stopped target sysinit.target - System Initialization. Nov 8 01:38:02.459888 systemd[1]: Stopped target local-fs.target - Local File Systems. Nov 8 01:38:02.479866 systemd[1]: Stopped target swap.target - Swaps. Nov 8 01:38:02.497748 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 8 01:38:02.498181 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Nov 8 01:38:02.524885 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Nov 8 01:38:02.545887 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 8 01:38:02.566635 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Nov 8 01:38:02.567087 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 8 01:38:02.588648 systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 8 01:38:02.589044 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Nov 8 01:38:02.620747 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Nov 8 01:38:02.621215 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Nov 8 01:38:02.640963 systemd[1]: Stopped target paths.target - Path Units. Nov 8 01:38:02.660620 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 8 01:38:02.661061 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 8 01:38:02.683778 systemd[1]: Stopped target slices.target - Slice Units. Nov 8 01:38:02.701765 systemd[1]: Stopped target sockets.target - Socket Units. Nov 8 01:38:02.721831 systemd[1]: iscsid.socket: Deactivated successfully. Nov 8 01:38:02.722136 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Nov 8 01:38:02.741894 systemd[1]: iscsiuio.socket: Deactivated successfully. Nov 8 01:38:02.742226 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 8 01:38:02.764966 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Nov 8 01:38:02.765399 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 8 01:38:02.784838 systemd[1]: ignition-files.service: Deactivated successfully. Nov 8 01:38:02.892329 ignition[1217]: INFO : Ignition 2.19.0 Nov 8 01:38:02.892329 ignition[1217]: INFO : Stage: umount Nov 8 01:38:02.892329 ignition[1217]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 8 01:38:02.892329 ignition[1217]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Nov 8 01:38:02.892329 ignition[1217]: INFO : umount: umount passed Nov 8 01:38:02.892329 ignition[1217]: INFO : POST message to Packet Timeline Nov 8 01:38:02.892329 ignition[1217]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Nov 8 01:38:02.785242 systemd[1]: Stopped ignition-files.service - Ignition (files). Nov 8 01:38:02.802847 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Nov 8 01:38:02.803261 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Nov 8 01:38:02.832423 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Nov 8 01:38:02.864407 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Nov 8 01:38:02.883355 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 8 01:38:02.883447 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Nov 8 01:38:02.903418 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 8 01:38:02.903517 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Nov 8 01:38:02.947284 systemd[1]: sysroot-boot.mount: Deactivated successfully. Nov 8 01:38:02.949180 systemd[1]: sysroot-boot.service: Deactivated successfully. Nov 8 01:38:02.949441 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Nov 8 01:38:02.964076 systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 8 01:38:02.964336 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Nov 8 01:38:03.872363 ignition[1217]: INFO : GET result: OK Nov 8 01:38:04.302260 ignition[1217]: INFO : Ignition finished successfully Nov 8 01:38:04.305340 systemd[1]: ignition-mount.service: Deactivated successfully. Nov 8 01:38:04.305647 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Nov 8 01:38:04.323400 systemd[1]: Stopped target network.target - Network. Nov 8 01:38:04.338392 systemd[1]: ignition-disks.service: Deactivated successfully. Nov 8 01:38:04.338568 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Nov 8 01:38:04.357491 systemd[1]: ignition-kargs.service: Deactivated successfully. Nov 8 01:38:04.357628 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Nov 8 01:38:04.377533 systemd[1]: ignition-setup.service: Deactivated successfully. Nov 8 01:38:04.377695 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Nov 8 01:38:04.396664 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Nov 8 01:38:04.396828 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Nov 8 01:38:04.415666 systemd[1]: initrd-setup-root.service: Deactivated successfully. Nov 8 01:38:04.415831 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Nov 8 01:38:04.435047 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Nov 8 01:38:04.451300 systemd-networkd[934]: enp1s0f0np0: DHCPv6 lease lost Nov 8 01:38:04.454625 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Nov 8 01:38:04.462369 systemd-networkd[934]: enp1s0f1np1: DHCPv6 lease lost Nov 8 01:38:04.473122 systemd[1]: systemd-resolved.service: Deactivated successfully. Nov 8 01:38:04.473439 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Nov 8 01:38:04.492671 systemd[1]: systemd-networkd.service: Deactivated successfully. Nov 8 01:38:04.492987 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Nov 8 01:38:04.513414 systemd[1]: systemd-networkd.socket: Deactivated successfully. Nov 8 01:38:04.513542 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Nov 8 01:38:04.545473 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Nov 8 01:38:04.552535 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Nov 8 01:38:04.552709 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 8 01:38:04.579578 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 8 01:38:04.579744 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 8 01:38:04.598572 systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 8 01:38:04.598743 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Nov 8 01:38:04.618557 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 8 01:38:04.618739 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 8 01:38:04.637820 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 8 01:38:04.659434 systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 8 01:38:04.659850 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 8 01:38:04.689707 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 8 01:38:04.689744 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Nov 8 01:38:04.715249 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 8 01:38:04.715275 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Nov 8 01:38:04.735356 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 8 01:38:04.735425 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Nov 8 01:38:04.778353 systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 8 01:38:04.778508 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Nov 8 01:38:04.816342 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 8 01:38:04.816495 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 8 01:38:04.870264 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Nov 8 01:38:04.899257 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 8 01:38:04.899411 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 8 01:38:04.920553 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Nov 8 01:38:04.920703 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 8 01:38:05.166378 systemd-journald[266]: Received SIGTERM from PID 1 (systemd). Nov 8 01:38:04.941540 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 8 01:38:04.941687 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Nov 8 01:38:04.962480 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 8 01:38:04.962652 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:38:04.985614 systemd[1]: network-cleanup.service: Deactivated successfully. Nov 8 01:38:04.985920 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Nov 8 01:38:05.003378 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 8 01:38:05.003641 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Nov 8 01:38:05.025625 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Nov 8 01:38:05.061630 systemd[1]: Starting initrd-switch-root.service - Switch Root... Nov 8 01:38:05.095653 systemd[1]: Switching root. Nov 8 01:38:05.270338 systemd-journald[266]: Journal stopped Nov 8 01:38:07.735569 kernel: SELinux: policy capability network_peer_controls=1 Nov 8 01:38:07.735584 kernel: SELinux: policy capability open_perms=1 Nov 8 01:38:07.735592 kernel: SELinux: policy capability extended_socket_class=1 Nov 8 01:38:07.735600 kernel: SELinux: policy capability always_check_network=0 Nov 8 01:38:07.735605 kernel: SELinux: policy capability cgroup_seclabel=1 Nov 8 01:38:07.735611 kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 8 01:38:07.735617 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Nov 8 01:38:07.735623 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Nov 8 01:38:07.735629 kernel: audit: type=1403 audit(1762565885.546:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Nov 8 01:38:07.735636 systemd[1]: Successfully loaded SELinux policy in 169.507ms. Nov 8 01:38:07.735644 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 7.263ms. Nov 8 01:38:07.735652 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Nov 8 01:38:07.735658 systemd[1]: Detected architecture x86-64. Nov 8 01:38:07.735665 systemd[1]: Detected first boot. Nov 8 01:38:07.735671 systemd[1]: Hostname set to . Nov 8 01:38:07.735680 systemd[1]: Initializing machine ID from random generator. Nov 8 01:38:07.735686 zram_generator::config[1283]: No configuration found. Nov 8 01:38:07.735694 systemd[1]: Populated /etc with preset unit settings. Nov 8 01:38:07.735700 systemd[1]: Queued start job for default target multi-user.target. Nov 8 01:38:07.735707 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Nov 8 01:38:07.735714 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Nov 8 01:38:07.735720 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Nov 8 01:38:07.735728 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Nov 8 01:38:07.735735 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Nov 8 01:38:07.735742 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Nov 8 01:38:07.735748 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Nov 8 01:38:07.735755 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Nov 8 01:38:07.735762 systemd[1]: Created slice user.slice - User and Session Slice. Nov 8 01:38:07.735769 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 8 01:38:07.735777 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 8 01:38:07.735784 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Nov 8 01:38:07.735790 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Nov 8 01:38:07.735798 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Nov 8 01:38:07.735805 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 8 01:38:07.735812 systemd[1]: Expecting device dev-ttyS1.device - /dev/ttyS1... Nov 8 01:38:07.735818 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 8 01:38:07.735825 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Nov 8 01:38:07.735833 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 8 01:38:07.735840 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 8 01:38:07.735847 systemd[1]: Reached target slices.target - Slice Units. Nov 8 01:38:07.735855 systemd[1]: Reached target swap.target - Swaps. Nov 8 01:38:07.735862 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Nov 8 01:38:07.735869 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Nov 8 01:38:07.735876 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 8 01:38:07.735884 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Nov 8 01:38:07.735892 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 8 01:38:07.735898 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 8 01:38:07.735905 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 8 01:38:07.735912 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Nov 8 01:38:07.735919 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Nov 8 01:38:07.735928 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Nov 8 01:38:07.735935 systemd[1]: Mounting media.mount - External Media Directory... Nov 8 01:38:07.735942 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:07.735949 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Nov 8 01:38:07.735956 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Nov 8 01:38:07.735963 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Nov 8 01:38:07.735971 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Nov 8 01:38:07.735979 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 8 01:38:07.735986 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 8 01:38:07.735994 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Nov 8 01:38:07.736000 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 8 01:38:07.736008 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 8 01:38:07.736015 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 8 01:38:07.736022 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Nov 8 01:38:07.736030 kernel: ACPI: bus type drm_connector registered Nov 8 01:38:07.736036 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 8 01:38:07.736044 kernel: fuse: init (API version 7.39) Nov 8 01:38:07.736051 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Nov 8 01:38:07.736059 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Nov 8 01:38:07.736066 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Nov 8 01:38:07.736073 kernel: loop: module loaded Nov 8 01:38:07.736080 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 8 01:38:07.736095 systemd-journald[1405]: Collecting audit messages is disabled. Nov 8 01:38:07.736111 systemd-journald[1405]: Journal started Nov 8 01:38:07.736126 systemd-journald[1405]: Runtime Journal (/run/log/journal/b48c9a95b3a3459e9d1312d762377d06) is 8.0M, max 639.9M, 631.9M free. Nov 8 01:38:07.765179 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 8 01:38:07.800207 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 8 01:38:07.835145 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Nov 8 01:38:07.869255 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 8 01:38:07.920189 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:07.941192 systemd[1]: Started systemd-journald.service - Journal Service. Nov 8 01:38:07.952063 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Nov 8 01:38:07.963440 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Nov 8 01:38:07.973415 systemd[1]: Mounted media.mount - External Media Directory. Nov 8 01:38:07.983424 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Nov 8 01:38:07.993382 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Nov 8 01:38:08.003388 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Nov 8 01:38:08.013551 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Nov 8 01:38:08.024633 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 8 01:38:08.035690 systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 8 01:38:08.035928 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Nov 8 01:38:08.048115 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 8 01:38:08.048587 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 8 01:38:08.060090 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 8 01:38:08.060561 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 8 01:38:08.071088 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 8 01:38:08.071649 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 8 01:38:08.083107 systemd[1]: modprobe@fuse.service: Deactivated successfully. Nov 8 01:38:08.083581 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Nov 8 01:38:08.094113 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 8 01:38:08.094651 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 8 01:38:08.104553 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 8 01:38:08.114506 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 8 01:38:08.125528 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Nov 8 01:38:08.136556 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 8 01:38:08.155215 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 8 01:38:08.186309 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Nov 8 01:38:08.197349 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Nov 8 01:38:08.207342 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Nov 8 01:38:08.209593 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Nov 8 01:38:08.220583 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Nov 8 01:38:08.231280 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 8 01:38:08.232023 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Nov 8 01:38:08.234609 systemd-journald[1405]: Time spent on flushing to /var/log/journal/b48c9a95b3a3459e9d1312d762377d06 is 12.925ms for 1357 entries. Nov 8 01:38:08.234609 systemd-journald[1405]: System Journal (/var/log/journal/b48c9a95b3a3459e9d1312d762377d06) is 8.0M, max 195.6M, 187.6M free. Nov 8 01:38:08.270938 systemd-journald[1405]: Received client request to flush runtime journal. Nov 8 01:38:08.250286 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 8 01:38:08.251250 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 8 01:38:08.257061 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 8 01:38:08.279152 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Nov 8 01:38:08.291533 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Nov 8 01:38:08.291920 systemd-tmpfiles[1443]: ACLs are not supported, ignoring. Nov 8 01:38:08.291930 systemd-tmpfiles[1443]: ACLs are not supported, ignoring. Nov 8 01:38:08.302410 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Nov 8 01:38:08.313427 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Nov 8 01:38:08.324394 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Nov 8 01:38:08.335402 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 8 01:38:08.345370 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 8 01:38:08.358779 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Nov 8 01:38:08.383327 systemd[1]: Starting systemd-sysusers.service - Create System Users... Nov 8 01:38:08.393530 udevadm[1450]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Nov 8 01:38:08.400429 systemd[1]: Finished systemd-sysusers.service - Create System Users. Nov 8 01:38:08.411349 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 8 01:38:08.429977 systemd-tmpfiles[1463]: ACLs are not supported, ignoring. Nov 8 01:38:08.429987 systemd-tmpfiles[1463]: ACLs are not supported, ignoring. Nov 8 01:38:08.432028 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 8 01:38:08.575523 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Nov 8 01:38:08.601425 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 8 01:38:08.614979 systemd-udevd[1469]: Using default interface naming scheme 'v255'. Nov 8 01:38:08.638604 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 8 01:38:08.654092 systemd[1]: Found device dev-ttyS1.device - /dev/ttyS1. Nov 8 01:38:08.668158 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input2 Nov 8 01:38:08.668210 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 35 scanned by (udev-worker) (1481) Nov 8 01:38:08.678152 kernel: ACPI: button: Sleep Button [SLPB] Nov 8 01:38:08.727686 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Micron_5300_MTFDDAK480TDT OEM. Nov 8 01:38:08.730818 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Nov 8 01:38:08.752205 kernel: IPMI message handler: version 39.2 Nov 8 01:38:08.752240 kernel: ACPI: button: Power Button [PWRF] Nov 8 01:38:08.813154 kernel: mousedev: PS/2 mouse device common for all mice Nov 8 01:38:08.813332 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 8 01:38:08.817149 kernel: ipmi device interface Nov 8 01:38:08.817176 kernel: i801_smbus 0000:00:1f.4: SPD Write Disable is set Nov 8 01:38:08.855277 kernel: i801_smbus 0000:00:1f.4: SMBus using PCI interrupt Nov 8 01:38:08.866020 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 8 01:38:08.885110 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Nov 8 01:38:08.891157 kernel: i2c i2c-0: 2/4 memory slots populated (from DMI) Nov 8 01:38:08.894147 kernel: ipmi_si: IPMI System Interface driver Nov 8 01:38:08.894176 kernel: mei_me 0000:00:16.4: Device doesn't have valid ME Interface Nov 8 01:38:08.894334 kernel: mei_me 0000:00:16.0: Device doesn't have valid ME Interface Nov 8 01:38:08.907805 kernel: ipmi_si dmi-ipmi-si.0: ipmi_platform: probing via SMBIOS Nov 8 01:38:08.907899 kernel: ipmi_platform: ipmi_si: SMBIOS: io 0xca2 regsize 1 spacing 1 irq 0 Nov 8 01:38:08.907910 kernel: ipmi_si: Adding SMBIOS-specified kcs state machine Nov 8 01:38:08.907918 kernel: ipmi_si IPI0001:00: ipmi_platform: probing via ACPI Nov 8 01:38:08.907995 kernel: ipmi_si IPI0001:00: ipmi_platform: [io 0x0ca2] regsize 1 spacing 1 irq 0 Nov 8 01:38:08.934152 kernel: iTCO_vendor_support: vendor-support=0 Nov 8 01:38:08.955147 kernel: ipmi_si dmi-ipmi-si.0: Removing SMBIOS-specified kcs state machine in favor of ACPI Nov 8 01:38:09.096959 kernel: ipmi_si: Adding ACPI-specified kcs state machine Nov 8 01:38:09.119373 kernel: ipmi_si: Trying ACPI-specified kcs state machine at i/o address 0xca2, slave address 0x20, irq 0 Nov 8 01:38:09.168277 kernel: iTCO_wdt iTCO_wdt: Found a Intel PCH TCO device (Version=6, TCOBASE=0x0400) Nov 8 01:38:09.168452 kernel: iTCO_wdt iTCO_wdt: initialized. heartbeat=30 sec (nowayout=0) Nov 8 01:38:09.168538 kernel: ipmi_si IPI0001:00: The BMC does not support clearing the recv irq bit, compensating, but the BMC needs to be fixed. Nov 8 01:38:09.171057 systemd[1]: Started systemd-userdbd.service - User Database Manager. Nov 8 01:38:09.252214 kernel: intel_rapl_common: Found RAPL domain package Nov 8 01:38:09.252250 kernel: ipmi_si IPI0001:00: IPMI message handler: Found new BMC (man_id: 0x002a7c, prod_id: 0x1b0f, dev_id: 0x20) Nov 8 01:38:09.252351 kernel: intel_rapl_common: Found RAPL domain core Nov 8 01:38:09.252362 kernel: intel_rapl_common: Found RAPL domain dram Nov 8 01:38:09.252405 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 8 01:38:09.345132 systemd-networkd[1553]: lo: Link UP Nov 8 01:38:09.345136 systemd-networkd[1553]: lo: Gained carrier Nov 8 01:38:09.347808 systemd-networkd[1553]: bond0: netdev ready Nov 8 01:38:09.348759 systemd-networkd[1553]: Enumeration completed Nov 8 01:38:09.348852 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 8 01:38:09.365427 systemd-networkd[1553]: enp1s0f0np0: Configuring with /etc/systemd/network/10-0c:42:a1:79:3d:90.network. Nov 8 01:38:09.375209 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Nov 8 01:38:09.405147 kernel: ipmi_si IPI0001:00: IPMI kcs interface initialized Nov 8 01:38:09.424148 kernel: ipmi_ssif: IPMI SSIF Interface driver Nov 8 01:38:09.425767 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Nov 8 01:38:09.450295 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Nov 8 01:38:09.458897 lvm[1587]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Nov 8 01:38:09.491876 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Nov 8 01:38:09.503368 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 8 01:38:09.520271 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Nov 8 01:38:09.522395 lvm[1590]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Nov 8 01:38:09.557862 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Nov 8 01:38:09.569348 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 8 01:38:09.581183 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Nov 8 01:38:09.581198 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 8 01:38:09.591180 systemd[1]: Reached target machines.target - Containers. Nov 8 01:38:09.600801 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Nov 8 01:38:09.623197 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Nov 8 01:38:09.635766 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Nov 8 01:38:09.646224 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 8 01:38:09.646687 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Nov 8 01:38:09.666579 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Nov 8 01:38:09.679002 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Nov 8 01:38:09.679554 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Nov 8 01:38:09.690866 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Nov 8 01:38:09.691347 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Nov 8 01:38:09.698836 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Nov 8 01:38:09.743164 kernel: loop0: detected capacity change from 0 to 140768 Nov 8 01:38:09.789179 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Nov 8 01:38:09.840184 kernel: loop1: detected capacity change from 0 to 224512 Nov 8 01:38:09.899149 kernel: loop2: detected capacity change from 0 to 8 Nov 8 01:38:09.943148 kernel: loop3: detected capacity change from 0 to 142488 Nov 8 01:38:09.978348 ldconfig[1596]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Nov 8 01:38:09.979679 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Nov 8 01:38:10.012176 kernel: loop4: detected capacity change from 0 to 140768 Nov 8 01:38:10.040174 kernel: loop5: detected capacity change from 0 to 224512 Nov 8 01:38:10.075676 kernel: loop6: detected capacity change from 0 to 8 Nov 8 01:38:10.075703 kernel: loop7: detected capacity change from 0 to 142488 Nov 8 01:38:10.096801 (sd-merge)[1616]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-packet'. Nov 8 01:38:10.097042 (sd-merge)[1616]: Merged extensions into '/usr'. Nov 8 01:38:10.099316 systemd[1]: Reloading requested from client PID 1600 ('systemd-sysext') (unit systemd-sysext.service)... Nov 8 01:38:10.099322 systemd[1]: Reloading... Nov 8 01:38:10.133191 zram_generator::config[1643]: No configuration found. Nov 8 01:38:10.196089 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 8 01:38:10.247807 systemd[1]: Reloading finished in 148 ms. Nov 8 01:38:10.260768 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Nov 8 01:38:10.284202 systemd[1]: Starting ensure-sysext.service... Nov 8 01:38:10.291879 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 8 01:38:10.304739 systemd[1]: Reloading requested from client PID 1704 ('systemctl') (unit ensure-sysext.service)... Nov 8 01:38:10.304746 systemd[1]: Reloading... Nov 8 01:38:10.313337 systemd-tmpfiles[1706]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Nov 8 01:38:10.313568 systemd-tmpfiles[1706]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Nov 8 01:38:10.314149 systemd-tmpfiles[1706]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Nov 8 01:38:10.314357 systemd-tmpfiles[1706]: ACLs are not supported, ignoring. Nov 8 01:38:10.314414 systemd-tmpfiles[1706]: ACLs are not supported, ignoring. Nov 8 01:38:10.318501 systemd-tmpfiles[1706]: Detected autofs mount point /boot during canonicalization of boot. Nov 8 01:38:10.318510 systemd-tmpfiles[1706]: Skipping /boot Nov 8 01:38:10.323423 systemd-tmpfiles[1706]: Detected autofs mount point /boot during canonicalization of boot. Nov 8 01:38:10.323429 systemd-tmpfiles[1706]: Skipping /boot Nov 8 01:38:10.336212 zram_generator::config[1735]: No configuration found. Nov 8 01:38:10.396986 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 8 01:38:10.450795 systemd[1]: Reloading finished in 145 ms. Nov 8 01:38:10.461098 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 8 01:38:10.476207 kernel: mlx5_core 0000:01:00.0 enp1s0f0np0: Link up Nov 8 01:38:10.499180 kernel: bond0: (slave enp1s0f0np0): Enslaving as a backup interface with an up link Nov 8 01:38:10.499781 systemd-networkd[1553]: enp1s0f1np1: Configuring with /etc/systemd/network/10-0c:42:a1:79:3d:91.network. Nov 8 01:38:10.514057 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:10.514936 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Nov 8 01:38:10.539800 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Nov 8 01:38:10.551319 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 8 01:38:10.551999 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 8 01:38:10.556074 augenrules[1818]: No rules Nov 8 01:38:10.561935 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 8 01:38:10.573884 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 8 01:38:10.583310 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 8 01:38:10.584117 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Nov 8 01:38:10.596388 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 8 01:38:10.606919 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Nov 8 01:38:10.618330 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:10.619327 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Nov 8 01:38:10.629625 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Nov 8 01:38:10.641640 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 8 01:38:10.641778 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 8 01:38:10.654760 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 8 01:38:10.654927 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 8 01:38:10.667113 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 8 01:38:10.667420 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 8 01:38:10.683680 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Nov 8 01:38:10.699221 kernel: mlx5_core 0000:01:00.1 enp1s0f1np1: Link up Nov 8 01:38:10.723794 systemd-networkd[1553]: bond0: Configuring with /etc/systemd/network/05-bond0.network. Nov 8 01:38:10.724188 kernel: bond0: (slave enp1s0f1np1): Enslaving as a backup interface with an up link Nov 8 01:38:10.725385 systemd-networkd[1553]: enp1s0f0np0: Link UP Nov 8 01:38:10.725716 systemd-networkd[1553]: enp1s0f0np0: Gained carrier Nov 8 01:38:10.726929 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:10.727223 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 8 01:38:10.744152 kernel: bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond Nov 8 01:38:10.751041 systemd-networkd[1553]: enp1s0f1np1: Reconfiguring with /etc/systemd/network/10-0c:42:a1:79:3d:90.network. Nov 8 01:38:10.751278 systemd-networkd[1553]: enp1s0f1np1: Link UP Nov 8 01:38:10.751351 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 8 01:38:10.751509 systemd-networkd[1553]: enp1s0f1np1: Gained carrier Nov 8 01:38:10.763571 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 8 01:38:10.769393 systemd-networkd[1553]: bond0: Link UP Nov 8 01:38:10.769662 systemd-networkd[1553]: bond0: Gained carrier Nov 8 01:38:10.775531 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 8 01:38:10.786265 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 8 01:38:10.787498 systemd[1]: Starting systemd-update-done.service - Update is Completed... Nov 8 01:38:10.797030 systemd-resolved[1831]: Positive Trust Anchors: Nov 8 01:38:10.797036 systemd-resolved[1831]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 8 01:38:10.797062 systemd-resolved[1831]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 8 01:38:10.798204 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Nov 8 01:38:10.798382 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:10.800185 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 8 01:38:10.800335 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 8 01:38:10.800562 systemd-resolved[1831]: Using system hostname 'ci-4081.3.6-n-3863d1ecf5'. Nov 8 01:38:10.812517 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 8 01:38:10.823434 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 8 01:38:10.823518 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 8 01:38:10.844494 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 8 01:38:10.844573 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 8 01:38:10.849678 kernel: bond0: (slave enp1s0f0np0): link status definitely up, 10000 Mbps full duplex Nov 8 01:38:10.849701 kernel: bond0: active interface up! Nov 8 01:38:10.860512 systemd[1]: Finished systemd-update-done.service - Update is Completed. Nov 8 01:38:10.872144 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Nov 8 01:38:10.886013 systemd[1]: Reached target network.target - Network. Nov 8 01:38:10.895233 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 8 01:38:10.907224 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:10.907362 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 8 01:38:10.917249 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 8 01:38:10.927785 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 8 01:38:10.937766 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 8 01:38:10.946896 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 8 01:38:10.957268 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 8 01:38:10.957342 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Nov 8 01:38:10.957395 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 8 01:38:10.958045 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 8 01:38:10.958124 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 8 01:38:10.978487 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 8 01:38:10.978566 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 8 01:38:10.982206 kernel: bond0: (slave enp1s0f1np1): link status definitely up, 10000 Mbps full duplex Nov 8 01:38:10.991463 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 8 01:38:10.991543 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 8 01:38:11.002442 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 8 01:38:11.002522 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 8 01:38:11.013208 systemd[1]: Finished ensure-sysext.service. Nov 8 01:38:11.022728 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 8 01:38:11.022759 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 8 01:38:11.037315 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Nov 8 01:38:11.082000 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Nov 8 01:38:11.094265 systemd[1]: Reached target sysinit.target - System Initialization. Nov 8 01:38:11.105218 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Nov 8 01:38:11.116189 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Nov 8 01:38:11.127177 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Nov 8 01:38:11.139180 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Nov 8 01:38:11.139193 systemd[1]: Reached target paths.target - Path Units. Nov 8 01:38:11.148170 systemd[1]: Reached target time-set.target - System Time Set. Nov 8 01:38:11.158239 systemd[1]: Started logrotate.timer - Daily rotation of log files. Nov 8 01:38:11.169211 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Nov 8 01:38:11.181170 systemd[1]: Reached target timers.target - Timer Units. Nov 8 01:38:11.190580 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Nov 8 01:38:11.201951 systemd[1]: Starting docker.socket - Docker Socket for the API... Nov 8 01:38:11.210630 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Nov 8 01:38:11.220414 systemd[1]: Listening on docker.socket - Docker Socket for the API. Nov 8 01:38:11.230193 systemd[1]: Reached target sockets.target - Socket Units. Nov 8 01:38:11.239173 systemd[1]: Reached target basic.target - Basic System. Nov 8 01:38:11.247250 systemd[1]: System is tainted: cgroupsv1 Nov 8 01:38:11.247270 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Nov 8 01:38:11.247282 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Nov 8 01:38:11.256168 systemd[1]: Starting containerd.service - containerd container runtime... Nov 8 01:38:11.265906 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Nov 8 01:38:11.275754 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Nov 8 01:38:11.284842 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Nov 8 01:38:11.288561 coreos-metadata[1879]: Nov 08 01:38:11.288 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Nov 8 01:38:11.294788 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Nov 8 01:38:11.296577 dbus-daemon[1880]: [system] SELinux support is enabled Nov 8 01:38:11.296810 jq[1883]: false Nov 8 01:38:11.304254 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Nov 8 01:38:11.304948 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Nov 8 01:38:11.313074 extend-filesystems[1885]: Found loop4 Nov 8 01:38:11.313074 extend-filesystems[1885]: Found loop5 Nov 8 01:38:11.313074 extend-filesystems[1885]: Found loop6 Nov 8 01:38:11.382254 kernel: EXT4-fs (sda9): resizing filesystem from 553472 to 116605649 blocks Nov 8 01:38:11.382272 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 35 scanned by (udev-worker) (1521) Nov 8 01:38:11.382293 extend-filesystems[1885]: Found loop7 Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda1 Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda2 Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda3 Nov 8 01:38:11.382293 extend-filesystems[1885]: Found usr Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda4 Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda6 Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda7 Nov 8 01:38:11.382293 extend-filesystems[1885]: Found sda9 Nov 8 01:38:11.382293 extend-filesystems[1885]: Checking size of /dev/sda9 Nov 8 01:38:11.382293 extend-filesystems[1885]: Resized partition /dev/sda9 Nov 8 01:38:11.314969 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Nov 8 01:38:11.515300 extend-filesystems[1897]: resize2fs 1.47.1 (20-May-2024) Nov 8 01:38:11.375277 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Nov 8 01:38:11.393875 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Nov 8 01:38:11.424759 systemd[1]: Starting systemd-logind.service - User Login Management... Nov 8 01:38:11.437084 systemd[1]: Starting tcsd.service - TCG Core Services Daemon... Nov 8 01:38:11.458882 systemd[1]: Starting update-engine.service - Update Engine... Nov 8 01:38:11.533689 update_engine[1914]: I20251108 01:38:11.499397 1914 main.cc:92] Flatcar Update Engine starting Nov 8 01:38:11.533689 update_engine[1914]: I20251108 01:38:11.500126 1914 update_check_scheduler.cc:74] Next update check in 2m43s Nov 8 01:38:11.475956 systemd-logind[1909]: Watching system buttons on /dev/input/event3 (Power Button) Nov 8 01:38:11.533967 jq[1915]: true Nov 8 01:38:11.475966 systemd-logind[1909]: Watching system buttons on /dev/input/event2 (Sleep Button) Nov 8 01:38:11.475976 systemd-logind[1909]: Watching system buttons on /dev/input/event0 (HID 0557:2419) Nov 8 01:38:11.476105 systemd-logind[1909]: New seat seat0. Nov 8 01:38:11.479922 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Nov 8 01:38:11.499552 systemd[1]: Started dbus.service - D-Bus System Message Bus. Nov 8 01:38:11.526464 systemd[1]: Started systemd-logind.service - User Login Management. Nov 8 01:38:11.554324 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Nov 8 01:38:11.554452 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Nov 8 01:38:11.554632 systemd[1]: motdgen.service: Deactivated successfully. Nov 8 01:38:11.554748 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Nov 8 01:38:11.564690 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Nov 8 01:38:11.564813 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Nov 8 01:38:11.571494 sshd_keygen[1912]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Nov 8 01:38:11.579255 (ntainerd)[1926]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Nov 8 01:38:11.580902 jq[1925]: true Nov 8 01:38:11.582738 dbus-daemon[1880]: [system] Successfully activated service 'org.freedesktop.systemd1' Nov 8 01:38:11.584191 tar[1919]: linux-amd64/LICENSE Nov 8 01:38:11.584378 tar[1919]: linux-amd64/helm Nov 8 01:38:11.585317 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Nov 8 01:38:11.595492 systemd[1]: tcsd.service: Skipped due to 'exec-condition'. Nov 8 01:38:11.595625 systemd[1]: Condition check resulted in tcsd.service - TCG Core Services Daemon being skipped. Nov 8 01:38:11.599268 systemd[1]: Started update-engine.service - Update Engine. Nov 8 01:38:11.620319 systemd[1]: Starting issuegen.service - Generate /run/issue... Nov 8 01:38:11.628281 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Nov 8 01:38:11.628385 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Nov 8 01:38:11.638813 bash[1958]: Updated "/home/core/.ssh/authorized_keys" Nov 8 01:38:11.639295 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Nov 8 01:38:11.639376 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Nov 8 01:38:11.650591 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Nov 8 01:38:11.651099 systemd[1]: Started locksmithd.service - Cluster reboot manager. Nov 8 01:38:11.680486 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Nov 8 01:38:11.691552 systemd[1]: issuegen.service: Deactivated successfully. Nov 8 01:38:11.691679 systemd[1]: Finished issuegen.service - Generate /run/issue. Nov 8 01:38:11.711804 locksmithd[1969]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Nov 8 01:38:11.713414 systemd[1]: Starting sshkeys.service... Nov 8 01:38:11.721051 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Nov 8 01:38:11.733192 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Nov 8 01:38:11.750254 containerd[1926]: time="2025-11-08T01:38:11.750210223Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Nov 8 01:38:11.754380 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Nov 8 01:38:11.763440 containerd[1926]: time="2025-11-08T01:38:11.763418140Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764202 containerd[1926]: time="2025-11-08T01:38:11.764185985Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.113-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764235 containerd[1926]: time="2025-11-08T01:38:11.764201415Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Nov 8 01:38:11.764235 containerd[1926]: time="2025-11-08T01:38:11.764210625Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Nov 8 01:38:11.764314 containerd[1926]: time="2025-11-08T01:38:11.764305369Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Nov 8 01:38:11.764335 containerd[1926]: time="2025-11-08T01:38:11.764316265Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764360 containerd[1926]: time="2025-11-08T01:38:11.764351067Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764383 containerd[1926]: time="2025-11-08T01:38:11.764359895Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764499 containerd[1926]: time="2025-11-08T01:38:11.764488626Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764520 containerd[1926]: time="2025-11-08T01:38:11.764498339Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764520 containerd[1926]: time="2025-11-08T01:38:11.764507087Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764520 containerd[1926]: time="2025-11-08T01:38:11.764512754Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764568 containerd[1926]: time="2025-11-08T01:38:11.764557903Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764690 containerd[1926]: time="2025-11-08T01:38:11.764681948Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764783 containerd[1926]: time="2025-11-08T01:38:11.764772434Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Nov 8 01:38:11.764783 containerd[1926]: time="2025-11-08T01:38:11.764782058Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Nov 8 01:38:11.764836 containerd[1926]: time="2025-11-08T01:38:11.764828562Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Nov 8 01:38:11.764870 containerd[1926]: time="2025-11-08T01:38:11.764861732Z" level=info msg="metadata content store policy set" policy=shared Nov 8 01:38:11.766836 coreos-metadata[1986]: Nov 08 01:38:11.765 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Nov 8 01:38:11.766634 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Nov 8 01:38:11.774633 containerd[1926]: time="2025-11-08T01:38:11.774618389Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Nov 8 01:38:11.774667 containerd[1926]: time="2025-11-08T01:38:11.774645121Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Nov 8 01:38:11.774667 containerd[1926]: time="2025-11-08T01:38:11.774655117Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Nov 8 01:38:11.774667 containerd[1926]: time="2025-11-08T01:38:11.774663905Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Nov 8 01:38:11.774725 containerd[1926]: time="2025-11-08T01:38:11.774673090Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Nov 8 01:38:11.774752 containerd[1926]: time="2025-11-08T01:38:11.774743976Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Nov 8 01:38:11.775281 containerd[1926]: time="2025-11-08T01:38:11.775198296Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Nov 8 01:38:11.775357 containerd[1926]: time="2025-11-08T01:38:11.775346442Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Nov 8 01:38:11.775382 containerd[1926]: time="2025-11-08T01:38:11.775359448Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Nov 8 01:38:11.775382 containerd[1926]: time="2025-11-08T01:38:11.775368034Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Nov 8 01:38:11.775382 containerd[1926]: time="2025-11-08T01:38:11.775377427Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775432 containerd[1926]: time="2025-11-08T01:38:11.775384889Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775432 containerd[1926]: time="2025-11-08T01:38:11.775391790Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775432 containerd[1926]: time="2025-11-08T01:38:11.775399424Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775432 containerd[1926]: time="2025-11-08T01:38:11.775408071Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775432 containerd[1926]: time="2025-11-08T01:38:11.775418481Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775432 containerd[1926]: time="2025-11-08T01:38:11.775425790Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775432 containerd[1926]: time="2025-11-08T01:38:11.775432211Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775444511Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775451805Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775458388Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775466973Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775473978Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775481562Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775488150Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775495195Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775502571Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775510573Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775516791Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775523208Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775529503Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775547 containerd[1926]: time="2025-11-08T01:38:11.775537866Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775551385Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775558946Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775565014Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775592575Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775602777Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775611692Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775618529Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775623963Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775632397Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775641404Z" level=info msg="NRI interface is disabled by configuration." Nov 8 01:38:11.775764 containerd[1926]: time="2025-11-08T01:38:11.775649500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Nov 8 01:38:11.775938 containerd[1926]: time="2025-11-08T01:38:11.775798360Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Nov 8 01:38:11.775938 containerd[1926]: time="2025-11-08T01:38:11.775830634Z" level=info msg="Connect containerd service" Nov 8 01:38:11.775938 containerd[1926]: time="2025-11-08T01:38:11.775852539Z" level=info msg="using legacy CRI server" Nov 8 01:38:11.775938 containerd[1926]: time="2025-11-08T01:38:11.775857569Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Nov 8 01:38:11.775938 containerd[1926]: time="2025-11-08T01:38:11.775918081Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Nov 8 01:38:11.776273 containerd[1926]: time="2025-11-08T01:38:11.776262568Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 8 01:38:11.776394 containerd[1926]: time="2025-11-08T01:38:11.776374629Z" level=info msg="Start subscribing containerd event" Nov 8 01:38:11.776421 containerd[1926]: time="2025-11-08T01:38:11.776403045Z" level=info msg="Start recovering state" Nov 8 01:38:11.776421 containerd[1926]: time="2025-11-08T01:38:11.776414160Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Nov 8 01:38:11.776466 containerd[1926]: time="2025-11-08T01:38:11.776438776Z" level=info msg=serving... address=/run/containerd/containerd.sock Nov 8 01:38:11.776490 containerd[1926]: time="2025-11-08T01:38:11.776438219Z" level=info msg="Start event monitor" Nov 8 01:38:11.776490 containerd[1926]: time="2025-11-08T01:38:11.776478268Z" level=info msg="Start snapshots syncer" Nov 8 01:38:11.776490 containerd[1926]: time="2025-11-08T01:38:11.776485972Z" level=info msg="Start cni network conf syncer for default" Nov 8 01:38:11.776562 containerd[1926]: time="2025-11-08T01:38:11.776490659Z" level=info msg="Start streaming server" Nov 8 01:38:11.776562 containerd[1926]: time="2025-11-08T01:38:11.776527276Z" level=info msg="containerd successfully booted in 0.026834s" Nov 8 01:38:11.778557 systemd[1]: Started containerd.service - containerd container runtime. Nov 8 01:38:11.809402 systemd[1]: Started getty@tty1.service - Getty on tty1. Nov 8 01:38:11.818032 systemd[1]: Started serial-getty@ttyS1.service - Serial Getty on ttyS1. Nov 8 01:38:11.827355 systemd[1]: Reached target getty.target - Login Prompts. Nov 8 01:38:11.871177 kernel: EXT4-fs (sda9): resized filesystem to 116605649 Nov 8 01:38:11.896092 extend-filesystems[1897]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Nov 8 01:38:11.896092 extend-filesystems[1897]: old_desc_blocks = 1, new_desc_blocks = 56 Nov 8 01:38:11.896092 extend-filesystems[1897]: The filesystem on /dev/sda9 is now 116605649 (4k) blocks long. Nov 8 01:38:11.926295 extend-filesystems[1885]: Resized filesystem in /dev/sda9 Nov 8 01:38:11.926295 extend-filesystems[1885]: Found sdb Nov 8 01:38:11.946280 tar[1919]: linux-amd64/README.md Nov 8 01:38:11.896567 systemd[1]: extend-filesystems.service: Deactivated successfully. Nov 8 01:38:11.896706 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Nov 8 01:38:11.956518 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Nov 8 01:38:12.641273 systemd-networkd[1553]: bond0: Gained IPv6LL Nov 8 01:38:12.642695 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Nov 8 01:38:12.653679 systemd[1]: Reached target network-online.target - Network is Online. Nov 8 01:38:12.682384 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 8 01:38:12.692928 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Nov 8 01:38:12.712550 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Nov 8 01:38:13.454921 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:13.478380 (kubelet)[2038]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 8 01:38:13.906277 kubelet[2038]: E1108 01:38:13.906216 2038 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 8 01:38:13.907605 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 8 01:38:13.907770 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 8 01:38:14.491734 systemd-timesyncd[1873]: Contacted time server 23.155.40.38:123 (0.flatcar.pool.ntp.org). Nov 8 01:38:14.491862 systemd-timesyncd[1873]: Initial clock synchronization to Sat 2025-11-08 01:38:14.385730 UTC. Nov 8 01:38:14.545018 kernel: mlx5_core 0000:01:00.0: lag map: port 1:1 port 2:2 Nov 8 01:38:14.545460 kernel: mlx5_core 0000:01:00.0: shared_fdb:0 mode:queue_affinity Nov 8 01:38:14.885509 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Nov 8 01:38:14.902407 systemd[1]: Started sshd@0-139.178.94.189:22-139.178.68.195:32792.service - OpenSSH per-connection server daemon (139.178.68.195:32792). Nov 8 01:38:14.941422 sshd[2059]: Accepted publickey for core from 139.178.68.195 port 32792 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:14.942593 sshd[2059]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:14.948197 systemd-logind[1909]: New session 1 of user core. Nov 8 01:38:14.949029 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Nov 8 01:38:14.965366 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Nov 8 01:38:14.978097 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Nov 8 01:38:15.002354 systemd[1]: Starting user@500.service - User Manager for UID 500... Nov 8 01:38:15.019475 (systemd)[2065]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Nov 8 01:38:15.092716 systemd[2065]: Queued start job for default target default.target. Nov 8 01:38:15.092955 systemd[2065]: Created slice app.slice - User Application Slice. Nov 8 01:38:15.092973 systemd[2065]: Reached target paths.target - Paths. Nov 8 01:38:15.092987 systemd[2065]: Reached target timers.target - Timers. Nov 8 01:38:15.104235 systemd[2065]: Starting dbus.socket - D-Bus User Message Bus Socket... Nov 8 01:38:15.107475 systemd[2065]: Listening on dbus.socket - D-Bus User Message Bus Socket. Nov 8 01:38:15.107514 systemd[2065]: Reached target sockets.target - Sockets. Nov 8 01:38:15.107527 systemd[2065]: Reached target basic.target - Basic System. Nov 8 01:38:15.107556 systemd[2065]: Reached target default.target - Main User Target. Nov 8 01:38:15.107578 systemd[2065]: Startup finished in 85ms. Nov 8 01:38:15.107659 systemd[1]: Started user@500.service - User Manager for UID 500. Nov 8 01:38:15.118026 systemd[1]: Started session-1.scope - Session 1 of User core. Nov 8 01:38:15.179314 systemd[1]: Started sshd@1-139.178.94.189:22-139.178.68.195:37594.service - OpenSSH per-connection server daemon (139.178.68.195:37594). Nov 8 01:38:15.205356 sshd[2078]: Accepted publickey for core from 139.178.68.195 port 37594 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:15.205980 sshd[2078]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:15.208288 systemd-logind[1909]: New session 2 of user core. Nov 8 01:38:15.220277 systemd[1]: Started session-2.scope - Session 2 of User core. Nov 8 01:38:15.267827 coreos-metadata[1986]: Nov 08 01:38:15.267 INFO Fetch successful Nov 8 01:38:15.274958 sshd[2078]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:15.281295 coreos-metadata[1879]: Nov 08 01:38:15.281 INFO Fetch successful Nov 8 01:38:15.288354 systemd[1]: Started sshd@2-139.178.94.189:22-139.178.68.195:37602.service - OpenSSH per-connection server daemon (139.178.68.195:37602). Nov 8 01:38:15.299639 systemd[1]: sshd@1-139.178.94.189:22-139.178.68.195:37594.service: Deactivated successfully. Nov 8 01:38:15.300224 unknown[1986]: wrote ssh authorized keys file for user: core Nov 8 01:38:15.300626 systemd[1]: session-2.scope: Deactivated successfully. Nov 8 01:38:15.300985 systemd-logind[1909]: Session 2 logged out. Waiting for processes to exit. Nov 8 01:38:15.301747 systemd-logind[1909]: Removed session 2. Nov 8 01:38:15.314319 update-ssh-keys[2090]: Updated "/home/core/.ssh/authorized_keys" Nov 8 01:38:15.314527 sshd[2084]: Accepted publickey for core from 139.178.68.195 port 37602 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:15.314723 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Nov 8 01:38:15.315353 sshd[2084]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:15.315685 systemd[1]: Finished sshkeys.service. Nov 8 01:38:15.355030 systemd-logind[1909]: New session 3 of user core. Nov 8 01:38:15.355614 systemd[1]: Started session-3.scope - Session 3 of User core. Nov 8 01:38:15.365530 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Nov 8 01:38:15.388308 systemd[1]: Starting packet-phone-home.service - Report Success to Packet... Nov 8 01:38:15.443576 sshd[2084]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:15.445026 systemd[1]: sshd@2-139.178.94.189:22-139.178.68.195:37602.service: Deactivated successfully. Nov 8 01:38:15.445892 systemd-logind[1909]: Session 3 logged out. Waiting for processes to exit. Nov 8 01:38:15.445907 systemd[1]: session-3.scope: Deactivated successfully. Nov 8 01:38:15.446365 systemd-logind[1909]: Removed session 3. Nov 8 01:38:15.748157 systemd[1]: Finished packet-phone-home.service - Report Success to Packet. Nov 8 01:38:15.760677 systemd[1]: Reached target multi-user.target - Multi-User System. Nov 8 01:38:15.770502 systemd[1]: Startup finished in 31.276s (kernel) + 10.392s (userspace) = 41.669s. Nov 8 01:38:15.790532 login[2005]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Nov 8 01:38:15.793955 systemd-logind[1909]: New session 4 of user core. Nov 8 01:38:15.794632 systemd[1]: Started session-4.scope - Session 4 of User core. Nov 8 01:38:15.803374 login[2004]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Nov 8 01:38:15.806098 systemd-logind[1909]: New session 5 of user core. Nov 8 01:38:15.806834 systemd[1]: Started session-5.scope - Session 5 of User core. Nov 8 01:38:24.159366 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Nov 8 01:38:24.172404 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 8 01:38:24.414612 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:24.419937 (kubelet)[2152]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 8 01:38:24.452469 kubelet[2152]: E1108 01:38:24.452414 2152 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 8 01:38:24.454488 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 8 01:38:24.454581 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 8 01:38:25.385548 systemd[1]: Started sshd@3-139.178.94.189:22-139.178.68.195:58028.service - OpenSSH per-connection server daemon (139.178.68.195:58028). Nov 8 01:38:25.413065 sshd[2171]: Accepted publickey for core from 139.178.68.195 port 58028 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:25.413668 sshd[2171]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:25.416395 systemd-logind[1909]: New session 6 of user core. Nov 8 01:38:25.425494 systemd[1]: Started session-6.scope - Session 6 of User core. Nov 8 01:38:25.476612 sshd[2171]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:25.487454 systemd[1]: Started sshd@4-139.178.94.189:22-139.178.68.195:58038.service - OpenSSH per-connection server daemon (139.178.68.195:58038). Nov 8 01:38:25.487799 systemd[1]: sshd@3-139.178.94.189:22-139.178.68.195:58028.service: Deactivated successfully. Nov 8 01:38:25.488664 systemd[1]: session-6.scope: Deactivated successfully. Nov 8 01:38:25.489125 systemd-logind[1909]: Session 6 logged out. Waiting for processes to exit. Nov 8 01:38:25.489974 systemd-logind[1909]: Removed session 6. Nov 8 01:38:25.514909 sshd[2177]: Accepted publickey for core from 139.178.68.195 port 58038 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:25.515599 sshd[2177]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:25.518240 systemd-logind[1909]: New session 7 of user core. Nov 8 01:38:25.532441 systemd[1]: Started session-7.scope - Session 7 of User core. Nov 8 01:38:25.581485 sshd[2177]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:25.594835 systemd[1]: Started sshd@5-139.178.94.189:22-139.178.68.195:58046.service - OpenSSH per-connection server daemon (139.178.68.195:58046). Nov 8 01:38:25.596120 systemd[1]: sshd@4-139.178.94.189:22-139.178.68.195:58038.service: Deactivated successfully. Nov 8 01:38:25.602189 systemd-logind[1909]: Session 7 logged out. Waiting for processes to exit. Nov 8 01:38:25.602803 systemd[1]: session-7.scope: Deactivated successfully. Nov 8 01:38:25.605432 systemd-logind[1909]: Removed session 7. Nov 8 01:38:25.656287 sshd[2184]: Accepted publickey for core from 139.178.68.195 port 58046 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:25.657049 sshd[2184]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:25.659771 systemd-logind[1909]: New session 8 of user core. Nov 8 01:38:25.673433 systemd[1]: Started session-8.scope - Session 8 of User core. Nov 8 01:38:25.724327 sshd[2184]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:25.744924 systemd[1]: Started sshd@6-139.178.94.189:22-139.178.68.195:58056.service - OpenSSH per-connection server daemon (139.178.68.195:58056). Nov 8 01:38:25.746272 systemd[1]: sshd@5-139.178.94.189:22-139.178.68.195:58046.service: Deactivated successfully. Nov 8 01:38:25.751889 systemd-logind[1909]: Session 8 logged out. Waiting for processes to exit. Nov 8 01:38:25.752478 systemd[1]: session-8.scope: Deactivated successfully. Nov 8 01:38:25.754927 systemd-logind[1909]: Removed session 8. Nov 8 01:38:25.799435 sshd[2192]: Accepted publickey for core from 139.178.68.195 port 58056 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:25.800054 sshd[2192]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:25.802731 systemd-logind[1909]: New session 9 of user core. Nov 8 01:38:25.812900 systemd[1]: Started session-9.scope - Session 9 of User core. Nov 8 01:38:25.884927 sudo[2199]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Nov 8 01:38:25.885079 sudo[2199]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 8 01:38:25.898857 sudo[2199]: pam_unix(sudo:session): session closed for user root Nov 8 01:38:25.899760 sshd[2192]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:25.916634 systemd[1]: Started sshd@7-139.178.94.189:22-139.178.68.195:58070.service - OpenSSH per-connection server daemon (139.178.68.195:58070). Nov 8 01:38:25.917050 systemd[1]: sshd@6-139.178.94.189:22-139.178.68.195:58056.service: Deactivated successfully. Nov 8 01:38:25.919065 systemd-logind[1909]: Session 9 logged out. Waiting for processes to exit. Nov 8 01:38:25.919226 systemd[1]: session-9.scope: Deactivated successfully. Nov 8 01:38:25.919972 systemd-logind[1909]: Removed session 9. Nov 8 01:38:25.964199 sshd[2201]: Accepted publickey for core from 139.178.68.195 port 58070 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:25.965174 sshd[2201]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:25.968736 systemd-logind[1909]: New session 10 of user core. Nov 8 01:38:25.983508 systemd[1]: Started session-10.scope - Session 10 of User core. Nov 8 01:38:26.041224 sudo[2209]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Nov 8 01:38:26.041430 sudo[2209]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 8 01:38:26.043555 sudo[2209]: pam_unix(sudo:session): session closed for user root Nov 8 01:38:26.046217 sudo[2208]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Nov 8 01:38:26.046368 sudo[2208]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 8 01:38:26.065446 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Nov 8 01:38:26.066689 auditctl[2212]: No rules Nov 8 01:38:26.066924 systemd[1]: audit-rules.service: Deactivated successfully. Nov 8 01:38:26.067089 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Nov 8 01:38:26.068919 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Nov 8 01:38:26.096874 augenrules[2231]: No rules Nov 8 01:38:26.097762 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Nov 8 01:38:26.098951 sudo[2208]: pam_unix(sudo:session): session closed for user root Nov 8 01:38:26.100730 sshd[2201]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:26.115582 systemd[1]: Started sshd@8-139.178.94.189:22-139.178.68.195:58078.service - OpenSSH per-connection server daemon (139.178.68.195:58078). Nov 8 01:38:26.116394 systemd[1]: sshd@7-139.178.94.189:22-139.178.68.195:58070.service: Deactivated successfully. Nov 8 01:38:26.118308 systemd[1]: session-10.scope: Deactivated successfully. Nov 8 01:38:26.119213 systemd-logind[1909]: Session 10 logged out. Waiting for processes to exit. Nov 8 01:38:26.120906 systemd-logind[1909]: Removed session 10. Nov 8 01:38:26.188504 sshd[2238]: Accepted publickey for core from 139.178.68.195 port 58078 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:38:26.189945 sshd[2238]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:38:26.194893 systemd-logind[1909]: New session 11 of user core. Nov 8 01:38:26.219832 systemd[1]: Started session-11.scope - Session 11 of User core. Nov 8 01:38:26.287915 sudo[2244]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Nov 8 01:38:26.288806 sudo[2244]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 8 01:38:26.643493 systemd[1]: Starting docker.service - Docker Application Container Engine... Nov 8 01:38:26.643598 (dockerd)[2270]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Nov 8 01:38:26.911627 dockerd[2270]: time="2025-11-08T01:38:26.911534903Z" level=info msg="Starting up" Nov 8 01:38:27.117689 dockerd[2270]: time="2025-11-08T01:38:27.117637437Z" level=info msg="Loading containers: start." Nov 8 01:38:27.208150 kernel: Initializing XFRM netlink socket Nov 8 01:38:27.314646 systemd-networkd[1553]: docker0: Link UP Nov 8 01:38:27.333068 dockerd[2270]: time="2025-11-08T01:38:27.333002578Z" level=info msg="Loading containers: done." Nov 8 01:38:27.345889 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1451761440-merged.mount: Deactivated successfully. Nov 8 01:38:27.346440 dockerd[2270]: time="2025-11-08T01:38:27.346395331Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Nov 8 01:38:27.346484 dockerd[2270]: time="2025-11-08T01:38:27.346443693Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Nov 8 01:38:27.346509 dockerd[2270]: time="2025-11-08T01:38:27.346500628Z" level=info msg="Daemon has completed initialization" Nov 8 01:38:27.361983 dockerd[2270]: time="2025-11-08T01:38:27.361953041Z" level=info msg="API listen on /run/docker.sock" Nov 8 01:38:27.361992 systemd[1]: Started docker.service - Docker Application Container Engine. Nov 8 01:38:28.448892 containerd[1926]: time="2025-11-08T01:38:28.448834327Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Nov 8 01:38:29.175474 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1616249429.mount: Deactivated successfully. Nov 8 01:38:30.372463 containerd[1926]: time="2025-11-08T01:38:30.372436286Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:30.372762 containerd[1926]: time="2025-11-08T01:38:30.372635214Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Nov 8 01:38:30.373117 containerd[1926]: time="2025-11-08T01:38:30.373105134Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:30.375117 containerd[1926]: time="2025-11-08T01:38:30.375104193Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:30.375689 containerd[1926]: time="2025-11-08T01:38:30.375674782Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 1.926815124s" Nov 8 01:38:30.375725 containerd[1926]: time="2025-11-08T01:38:30.375692672Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Nov 8 01:38:30.376049 containerd[1926]: time="2025-11-08T01:38:30.376022002Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Nov 8 01:38:31.893175 containerd[1926]: time="2025-11-08T01:38:31.893126199Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:31.893418 containerd[1926]: time="2025-11-08T01:38:31.893301186Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Nov 8 01:38:31.893719 containerd[1926]: time="2025-11-08T01:38:31.893701955Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:31.895772 containerd[1926]: time="2025-11-08T01:38:31.895755553Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:31.896321 containerd[1926]: time="2025-11-08T01:38:31.896305651Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 1.520252684s" Nov 8 01:38:31.896367 containerd[1926]: time="2025-11-08T01:38:31.896322256Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Nov 8 01:38:31.896561 containerd[1926]: time="2025-11-08T01:38:31.896548706Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Nov 8 01:38:32.995655 containerd[1926]: time="2025-11-08T01:38:32.995626022Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:32.995886 containerd[1926]: time="2025-11-08T01:38:32.995852624Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Nov 8 01:38:32.996313 containerd[1926]: time="2025-11-08T01:38:32.996271622Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:32.997953 containerd[1926]: time="2025-11-08T01:38:32.997913477Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:32.998624 containerd[1926]: time="2025-11-08T01:38:32.998581211Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 1.10201638s" Nov 8 01:38:32.998624 containerd[1926]: time="2025-11-08T01:38:32.998598231Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Nov 8 01:38:32.998871 containerd[1926]: time="2025-11-08T01:38:32.998832321Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Nov 8 01:38:33.934861 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2395936542.mount: Deactivated successfully. Nov 8 01:38:34.188591 containerd[1926]: time="2025-11-08T01:38:34.188524751Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:34.188865 containerd[1926]: time="2025-11-08T01:38:34.188709838Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Nov 8 01:38:34.189131 containerd[1926]: time="2025-11-08T01:38:34.189118341Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:34.190047 containerd[1926]: time="2025-11-08T01:38:34.190034957Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:34.190478 containerd[1926]: time="2025-11-08T01:38:34.190439031Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 1.191589125s" Nov 8 01:38:34.190478 containerd[1926]: time="2025-11-08T01:38:34.190455246Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Nov 8 01:38:34.190715 containerd[1926]: time="2025-11-08T01:38:34.190682703Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Nov 8 01:38:34.531021 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Nov 8 01:38:34.548502 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 8 01:38:34.789892 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:34.792252 (kubelet)[2527]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 8 01:38:34.815983 kubelet[2527]: E1108 01:38:34.815940 2527 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 8 01:38:34.817076 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 8 01:38:34.817173 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 8 01:38:34.865078 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3237620213.mount: Deactivated successfully. Nov 8 01:38:35.419867 containerd[1926]: time="2025-11-08T01:38:35.419812209Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:35.420106 containerd[1926]: time="2025-11-08T01:38:35.420007600Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Nov 8 01:38:35.420483 containerd[1926]: time="2025-11-08T01:38:35.420443658Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:35.422047 containerd[1926]: time="2025-11-08T01:38:35.422032287Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:35.422731 containerd[1926]: time="2025-11-08T01:38:35.422692547Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.23199468s" Nov 8 01:38:35.422731 containerd[1926]: time="2025-11-08T01:38:35.422710718Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Nov 8 01:38:35.422945 containerd[1926]: time="2025-11-08T01:38:35.422934908Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Nov 8 01:38:35.989419 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1521445440.mount: Deactivated successfully. Nov 8 01:38:35.990659 containerd[1926]: time="2025-11-08T01:38:35.990642154Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:35.990874 containerd[1926]: time="2025-11-08T01:38:35.990854478Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Nov 8 01:38:35.991137 containerd[1926]: time="2025-11-08T01:38:35.991125350Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:35.992454 containerd[1926]: time="2025-11-08T01:38:35.992441454Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:35.992909 containerd[1926]: time="2025-11-08T01:38:35.992897045Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 569.946858ms" Nov 8 01:38:35.992942 containerd[1926]: time="2025-11-08T01:38:35.992912062Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Nov 8 01:38:35.993218 containerd[1926]: time="2025-11-08T01:38:35.993191571Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Nov 8 01:38:36.611116 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3276178094.mount: Deactivated successfully. Nov 8 01:38:38.420864 containerd[1926]: time="2025-11-08T01:38:38.420838527Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:38.421107 containerd[1926]: time="2025-11-08T01:38:38.421018996Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Nov 8 01:38:38.421571 containerd[1926]: time="2025-11-08T01:38:38.421531642Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:38.423798 containerd[1926]: time="2025-11-08T01:38:38.423756863Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:38.424416 containerd[1926]: time="2025-11-08T01:38:38.424375665Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.4311521s" Nov 8 01:38:38.424416 containerd[1926]: time="2025-11-08T01:38:38.424391651Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Nov 8 01:38:40.151342 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:40.167514 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 8 01:38:40.182068 systemd[1]: Reloading requested from client PID 2712 ('systemctl') (unit session-11.scope)... Nov 8 01:38:40.182075 systemd[1]: Reloading... Nov 8 01:38:40.212234 zram_generator::config[2751]: No configuration found. Nov 8 01:38:40.291775 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 8 01:38:40.351945 systemd[1]: Reloading finished in 169 ms. Nov 8 01:38:40.398145 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Nov 8 01:38:40.398222 systemd[1]: kubelet.service: Failed with result 'signal'. Nov 8 01:38:40.398372 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:40.399611 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 8 01:38:40.644275 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:40.646605 (kubelet)[2828]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 8 01:38:40.666992 kubelet[2828]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 8 01:38:40.666992 kubelet[2828]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 8 01:38:40.666992 kubelet[2828]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 8 01:38:40.667272 kubelet[2828]: I1108 01:38:40.666999 2828 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 8 01:38:40.959551 kubelet[2828]: I1108 01:38:40.959478 2828 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 8 01:38:40.959551 kubelet[2828]: I1108 01:38:40.959500 2828 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 8 01:38:40.960335 kubelet[2828]: I1108 01:38:40.960291 2828 server.go:954] "Client rotation is on, will bootstrap in background" Nov 8 01:38:40.998793 kubelet[2828]: E1108 01:38:40.998777 2828 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://139.178.94.189:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 139.178.94.189:6443: connect: connection refused" logger="UnhandledError" Nov 8 01:38:40.999479 kubelet[2828]: I1108 01:38:40.999464 2828 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 8 01:38:41.003954 kubelet[2828]: E1108 01:38:41.003939 2828 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Nov 8 01:38:41.003981 kubelet[2828]: I1108 01:38:41.003957 2828 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Nov 8 01:38:41.012799 kubelet[2828]: I1108 01:38:41.012784 2828 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 8 01:38:41.014053 kubelet[2828]: I1108 01:38:41.014009 2828 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 8 01:38:41.014277 kubelet[2828]: I1108 01:38:41.014034 2828 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.6-n-3863d1ecf5","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Nov 8 01:38:41.014277 kubelet[2828]: I1108 01:38:41.014255 2828 topology_manager.go:138] "Creating topology manager with none policy" Nov 8 01:38:41.014277 kubelet[2828]: I1108 01:38:41.014262 2828 container_manager_linux.go:304] "Creating device plugin manager" Nov 8 01:38:41.014386 kubelet[2828]: I1108 01:38:41.014329 2828 state_mem.go:36] "Initialized new in-memory state store" Nov 8 01:38:41.017534 kubelet[2828]: I1108 01:38:41.017499 2828 kubelet.go:446] "Attempting to sync node with API server" Nov 8 01:38:41.017534 kubelet[2828]: I1108 01:38:41.017513 2828 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 8 01:38:41.017534 kubelet[2828]: I1108 01:38:41.017523 2828 kubelet.go:352] "Adding apiserver pod source" Nov 8 01:38:41.017534 kubelet[2828]: I1108 01:38:41.017528 2828 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 8 01:38:41.020119 kubelet[2828]: I1108 01:38:41.020111 2828 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Nov 8 01:38:41.020428 kubelet[2828]: I1108 01:38:41.020421 2828 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 8 01:38:41.020998 kubelet[2828]: W1108 01:38:41.020990 2828 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Nov 8 01:38:41.021849 kubelet[2828]: W1108 01:38:41.021805 2828 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.94.189:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 139.178.94.189:6443: connect: connection refused Nov 8 01:38:41.021849 kubelet[2828]: W1108 01:38:41.021801 2828 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.94.189:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-3863d1ecf5&limit=500&resourceVersion=0": dial tcp 139.178.94.189:6443: connect: connection refused Nov 8 01:38:41.021938 kubelet[2828]: E1108 01:38:41.021848 2828 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://139.178.94.189:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.94.189:6443: connect: connection refused" logger="UnhandledError" Nov 8 01:38:41.021938 kubelet[2828]: E1108 01:38:41.021850 2828 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://139.178.94.189:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-3863d1ecf5&limit=500&resourceVersion=0\": dial tcp 139.178.94.189:6443: connect: connection refused" logger="UnhandledError" Nov 8 01:38:41.022728 kubelet[2828]: I1108 01:38:41.022693 2828 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 8 01:38:41.022728 kubelet[2828]: I1108 01:38:41.022724 2828 server.go:1287] "Started kubelet" Nov 8 01:38:41.022843 kubelet[2828]: I1108 01:38:41.022827 2828 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 8 01:38:41.023866 kubelet[2828]: I1108 01:38:41.023832 2828 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 8 01:38:41.023866 kubelet[2828]: I1108 01:38:41.023863 2828 server.go:479] "Adding debug handlers to kubelet server" Nov 8 01:38:41.023935 kubelet[2828]: I1108 01:38:41.023919 2828 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 8 01:38:41.023968 kubelet[2828]: I1108 01:38:41.023948 2828 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 8 01:38:41.023999 kubelet[2828]: I1108 01:38:41.023970 2828 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 8 01:38:41.024027 kubelet[2828]: E1108 01:38:41.024003 2828 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.6-n-3863d1ecf5\" not found" Nov 8 01:38:41.024027 kubelet[2828]: I1108 01:38:41.024017 2828 reconciler.go:26] "Reconciler: start to sync state" Nov 8 01:38:41.024101 kubelet[2828]: E1108 01:38:41.024082 2828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.94.189:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-3863d1ecf5?timeout=10s\": dial tcp 139.178.94.189:6443: connect: connection refused" interval="200ms" Nov 8 01:38:41.024215 kubelet[2828]: W1108 01:38:41.024189 2828 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.94.189:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.94.189:6443: connect: connection refused Nov 8 01:38:41.024269 kubelet[2828]: E1108 01:38:41.024223 2828 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://139.178.94.189:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.94.189:6443: connect: connection refused" logger="UnhandledError" Nov 8 01:38:41.024303 kubelet[2828]: I1108 01:38:41.024296 2828 factory.go:221] Registration of the systemd container factory successfully Nov 8 01:38:41.024360 kubelet[2828]: I1108 01:38:41.024346 2828 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 8 01:38:41.026000 kubelet[2828]: I1108 01:38:41.024536 2828 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 8 01:38:41.026194 kubelet[2828]: I1108 01:38:41.026183 2828 factory.go:221] Registration of the containerd container factory successfully Nov 8 01:38:41.026241 kubelet[2828]: I1108 01:38:41.026229 2828 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 8 01:38:41.026536 kubelet[2828]: E1108 01:38:41.026522 2828 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 8 01:38:41.028366 kubelet[2828]: E1108 01:38:41.027341 2828 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.94.189:6443/api/v1/namespaces/default/events\": dial tcp 139.178.94.189:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.6-n-3863d1ecf5.1875e45ddfc468ca default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.6-n-3863d1ecf5,UID:ci-4081.3.6-n-3863d1ecf5,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.6-n-3863d1ecf5,},FirstTimestamp:2025-11-08 01:38:41.022699722 +0000 UTC m=+0.374280407,LastTimestamp:2025-11-08 01:38:41.022699722 +0000 UTC m=+0.374280407,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.6-n-3863d1ecf5,}" Nov 8 01:38:41.033238 kubelet[2828]: I1108 01:38:41.033213 2828 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 8 01:38:41.033774 kubelet[2828]: I1108 01:38:41.033749 2828 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 8 01:38:41.033774 kubelet[2828]: I1108 01:38:41.033776 2828 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 8 01:38:41.033843 kubelet[2828]: I1108 01:38:41.033804 2828 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 8 01:38:41.033843 kubelet[2828]: I1108 01:38:41.033808 2828 kubelet.go:2382] "Starting kubelet main sync loop" Nov 8 01:38:41.033843 kubelet[2828]: E1108 01:38:41.033833 2828 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 8 01:38:41.034962 kubelet[2828]: W1108 01:38:41.034951 2828 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.94.189:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.94.189:6443: connect: connection refused Nov 8 01:38:41.034997 kubelet[2828]: E1108 01:38:41.034974 2828 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.94.189:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.94.189:6443: connect: connection refused" logger="UnhandledError" Nov 8 01:38:41.095602 kubelet[2828]: I1108 01:38:41.095576 2828 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 8 01:38:41.095602 kubelet[2828]: I1108 01:38:41.095595 2828 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 8 01:38:41.095807 kubelet[2828]: I1108 01:38:41.095633 2828 state_mem.go:36] "Initialized new in-memory state store" Nov 8 01:38:41.096926 kubelet[2828]: I1108 01:38:41.096913 2828 policy_none.go:49] "None policy: Start" Nov 8 01:38:41.096926 kubelet[2828]: I1108 01:38:41.096928 2828 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 8 01:38:41.097012 kubelet[2828]: I1108 01:38:41.096940 2828 state_mem.go:35] "Initializing new in-memory state store" Nov 8 01:38:41.099359 kubelet[2828]: I1108 01:38:41.099352 2828 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 8 01:38:41.099497 kubelet[2828]: I1108 01:38:41.099439 2828 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 8 01:38:41.099497 kubelet[2828]: I1108 01:38:41.099447 2828 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 8 01:38:41.099544 kubelet[2828]: I1108 01:38:41.099528 2828 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 8 01:38:41.099741 kubelet[2828]: E1108 01:38:41.099734 2828 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 8 01:38:41.099766 kubelet[2828]: E1108 01:38:41.099751 2828 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.6-n-3863d1ecf5\" not found" Nov 8 01:38:41.141694 kubelet[2828]: E1108 01:38:41.141657 2828 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-3863d1ecf5\" not found" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.143839 kubelet[2828]: E1108 01:38:41.143799 2828 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-3863d1ecf5\" not found" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.146731 kubelet[2828]: E1108 01:38:41.146690 2828 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-3863d1ecf5\" not found" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.203395 kubelet[2828]: I1108 01:38:41.203302 2828 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.204090 kubelet[2828]: E1108 01:38:41.203991 2828 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.94.189:6443/api/v1/nodes\": dial tcp 139.178.94.189:6443: connect: connection refused" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.225361 kubelet[2828]: E1108 01:38:41.225091 2828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.94.189:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-3863d1ecf5?timeout=10s\": dial tcp 139.178.94.189:6443: connect: connection refused" interval="400ms" Nov 8 01:38:41.324702 kubelet[2828]: I1108 01:38:41.324629 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7922af8ce0cb5d65be29beff962b809e-k8s-certs\") pod \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" (UID: \"7922af8ce0cb5d65be29beff962b809e\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.324702 kubelet[2828]: I1108 01:38:41.324679 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.324869 kubelet[2828]: I1108 01:38:41.324716 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.324869 kubelet[2828]: I1108 01:38:41.324741 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/67df4e4b57fbbb4d0e729ac8bd6036d6-kubeconfig\") pod \"kube-scheduler-ci-4081.3.6-n-3863d1ecf5\" (UID: \"67df4e4b57fbbb4d0e729ac8bd6036d6\") " pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.324869 kubelet[2828]: I1108 01:38:41.324765 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7922af8ce0cb5d65be29beff962b809e-ca-certs\") pod \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" (UID: \"7922af8ce0cb5d65be29beff962b809e\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.324869 kubelet[2828]: I1108 01:38:41.324793 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7922af8ce0cb5d65be29beff962b809e-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" (UID: \"7922af8ce0cb5d65be29beff962b809e\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.324869 kubelet[2828]: I1108 01:38:41.324816 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-ca-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.325087 kubelet[2828]: I1108 01:38:41.324836 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.325087 kubelet[2828]: I1108 01:38:41.324865 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.409303 kubelet[2828]: I1108 01:38:41.409198 2828 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.410010 kubelet[2828]: E1108 01:38:41.409890 2828 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.94.189:6443/api/v1/nodes\": dial tcp 139.178.94.189:6443: connect: connection refused" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.444426 containerd[1926]: time="2025-11-08T01:38:41.444278017Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.6-n-3863d1ecf5,Uid:67df4e4b57fbbb4d0e729ac8bd6036d6,Namespace:kube-system,Attempt:0,}" Nov 8 01:38:41.445734 containerd[1926]: time="2025-11-08T01:38:41.445669343Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.6-n-3863d1ecf5,Uid:7922af8ce0cb5d65be29beff962b809e,Namespace:kube-system,Attempt:0,}" Nov 8 01:38:41.448074 containerd[1926]: time="2025-11-08T01:38:41.448037356Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.6-n-3863d1ecf5,Uid:332b2243d2a6139435e1aec31ab74d79,Namespace:kube-system,Attempt:0,}" Nov 8 01:38:41.626186 kubelet[2828]: E1108 01:38:41.626121 2828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.94.189:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-3863d1ecf5?timeout=10s\": dial tcp 139.178.94.189:6443: connect: connection refused" interval="800ms" Nov 8 01:38:41.811704 kubelet[2828]: I1108 01:38:41.811662 2828 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.811976 kubelet[2828]: E1108 01:38:41.811902 2828 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.94.189:6443/api/v1/nodes\": dial tcp 139.178.94.189:6443: connect: connection refused" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:41.988383 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1765164052.mount: Deactivated successfully. Nov 8 01:38:41.989901 containerd[1926]: time="2025-11-08T01:38:41.989858223Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 8 01:38:41.990470 containerd[1926]: time="2025-11-08T01:38:41.990425278Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 8 01:38:41.990832 containerd[1926]: time="2025-11-08T01:38:41.990773966Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Nov 8 01:38:41.990938 containerd[1926]: time="2025-11-08T01:38:41.990907193Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 8 01:38:41.991195 containerd[1926]: time="2025-11-08T01:38:41.991126610Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Nov 8 01:38:41.991359 containerd[1926]: time="2025-11-08T01:38:41.991313457Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Nov 8 01:38:41.991718 containerd[1926]: time="2025-11-08T01:38:41.991677894Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 8 01:38:41.993853 containerd[1926]: time="2025-11-08T01:38:41.993811953Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 8 01:38:41.994357 containerd[1926]: time="2025-11-08T01:38:41.994315932Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 548.569582ms" Nov 8 01:38:41.994717 containerd[1926]: time="2025-11-08T01:38:41.994677340Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 550.223753ms" Nov 8 01:38:41.995757 containerd[1926]: time="2025-11-08T01:38:41.995715243Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 547.634132ms" Nov 8 01:38:42.083298 kubelet[2828]: W1108 01:38:42.083235 2828 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.94.189:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.94.189:6443: connect: connection refused Nov 8 01:38:42.083298 kubelet[2828]: E1108 01:38:42.083281 2828 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.94.189:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.94.189:6443: connect: connection refused" logger="UnhandledError" Nov 8 01:38:42.101085 containerd[1926]: time="2025-11-08T01:38:42.101036487Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:38:42.101085 containerd[1926]: time="2025-11-08T01:38:42.101073421Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:38:42.101197 containerd[1926]: time="2025-11-08T01:38:42.101088905Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:42.101396 containerd[1926]: time="2025-11-08T01:38:42.101341300Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:42.101766 containerd[1926]: time="2025-11-08T01:38:42.101738302Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:38:42.101796 containerd[1926]: time="2025-11-08T01:38:42.101763919Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:38:42.101796 containerd[1926]: time="2025-11-08T01:38:42.101771280Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:42.101796 containerd[1926]: time="2025-11-08T01:38:42.101574563Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:38:42.101848 containerd[1926]: time="2025-11-08T01:38:42.101792881Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:38:42.101848 containerd[1926]: time="2025-11-08T01:38:42.101803202Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:42.101848 containerd[1926]: time="2025-11-08T01:38:42.101817170Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:42.101900 containerd[1926]: time="2025-11-08T01:38:42.101849898Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:42.150642 containerd[1926]: time="2025-11-08T01:38:42.150619881Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.6-n-3863d1ecf5,Uid:67df4e4b57fbbb4d0e729ac8bd6036d6,Namespace:kube-system,Attempt:0,} returns sandbox id \"7c877c7bcdc85bea71928cde0ed3d191a46150136182a1e47ea1798f6364328e\"" Nov 8 01:38:42.151469 containerd[1926]: time="2025-11-08T01:38:42.151454039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.6-n-3863d1ecf5,Uid:332b2243d2a6139435e1aec31ab74d79,Namespace:kube-system,Attempt:0,} returns sandbox id \"7ca6abb59ed7ace6d3a73f09390bec3ae13edbdb7d06a4c48884b8faea8100f8\"" Nov 8 01:38:42.151579 containerd[1926]: time="2025-11-08T01:38:42.151569547Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.6-n-3863d1ecf5,Uid:7922af8ce0cb5d65be29beff962b809e,Namespace:kube-system,Attempt:0,} returns sandbox id \"286849cfb302416f0f4c998f4cda594a504a8133023b2706f0acf83ffaf67cfd\"" Nov 8 01:38:42.152099 containerd[1926]: time="2025-11-08T01:38:42.152083713Z" level=info msg="CreateContainer within sandbox \"7c877c7bcdc85bea71928cde0ed3d191a46150136182a1e47ea1798f6364328e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Nov 8 01:38:42.152264 containerd[1926]: time="2025-11-08T01:38:42.152253738Z" level=info msg="CreateContainer within sandbox \"7ca6abb59ed7ace6d3a73f09390bec3ae13edbdb7d06a4c48884b8faea8100f8\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Nov 8 01:38:42.152299 containerd[1926]: time="2025-11-08T01:38:42.152287917Z" level=info msg="CreateContainer within sandbox \"286849cfb302416f0f4c998f4cda594a504a8133023b2706f0acf83ffaf67cfd\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Nov 8 01:38:42.158091 containerd[1926]: time="2025-11-08T01:38:42.158049443Z" level=info msg="CreateContainer within sandbox \"286849cfb302416f0f4c998f4cda594a504a8133023b2706f0acf83ffaf67cfd\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"d6b584ec876711a2657869e8c55a9538c4048cb56686d4f0bac5b4a4029e6ce1\"" Nov 8 01:38:42.158301 containerd[1926]: time="2025-11-08T01:38:42.158270071Z" level=info msg="StartContainer for \"d6b584ec876711a2657869e8c55a9538c4048cb56686d4f0bac5b4a4029e6ce1\"" Nov 8 01:38:42.158621 containerd[1926]: time="2025-11-08T01:38:42.158579663Z" level=info msg="CreateContainer within sandbox \"7c877c7bcdc85bea71928cde0ed3d191a46150136182a1e47ea1798f6364328e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"7172fd6014d9c01cf9ec4650fc83bbe3b34872780e52c686b18e0ef8bb6644d5\"" Nov 8 01:38:42.158801 containerd[1926]: time="2025-11-08T01:38:42.158760814Z" level=info msg="StartContainer for \"7172fd6014d9c01cf9ec4650fc83bbe3b34872780e52c686b18e0ef8bb6644d5\"" Nov 8 01:38:42.160364 containerd[1926]: time="2025-11-08T01:38:42.160316204Z" level=info msg="CreateContainer within sandbox \"7ca6abb59ed7ace6d3a73f09390bec3ae13edbdb7d06a4c48884b8faea8100f8\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f399ef576054abdc112bd6fa6dde74fc37faa822c5907a82bcb4b2a6a92f201d\"" Nov 8 01:38:42.160575 containerd[1926]: time="2025-11-08T01:38:42.160536433Z" level=info msg="StartContainer for \"f399ef576054abdc112bd6fa6dde74fc37faa822c5907a82bcb4b2a6a92f201d\"" Nov 8 01:38:42.207843 containerd[1926]: time="2025-11-08T01:38:42.207809601Z" level=info msg="StartContainer for \"d6b584ec876711a2657869e8c55a9538c4048cb56686d4f0bac5b4a4029e6ce1\" returns successfully" Nov 8 01:38:42.207843 containerd[1926]: time="2025-11-08T01:38:42.207809922Z" level=info msg="StartContainer for \"f399ef576054abdc112bd6fa6dde74fc37faa822c5907a82bcb4b2a6a92f201d\" returns successfully" Nov 8 01:38:42.207963 containerd[1926]: time="2025-11-08T01:38:42.207816855Z" level=info msg="StartContainer for \"7172fd6014d9c01cf9ec4650fc83bbe3b34872780e52c686b18e0ef8bb6644d5\" returns successfully" Nov 8 01:38:42.613931 kubelet[2828]: I1108 01:38:42.613911 2828 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.687954 kubelet[2828]: E1108 01:38:42.687932 2828 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.6-n-3863d1ecf5\" not found" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.786819 kubelet[2828]: I1108 01:38:42.786796 2828 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.824738 kubelet[2828]: I1108 01:38:42.824673 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.830930 kubelet[2828]: E1108 01:38:42.830857 2828 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-3863d1ecf5\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.830930 kubelet[2828]: I1108 01:38:42.830898 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.833162 kubelet[2828]: E1108 01:38:42.833118 2828 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.833162 kubelet[2828]: I1108 01:38:42.833166 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:42.835227 kubelet[2828]: E1108 01:38:42.835166 2828 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:43.019589 kubelet[2828]: I1108 01:38:43.019399 2828 apiserver.go:52] "Watching apiserver" Nov 8 01:38:43.024200 kubelet[2828]: I1108 01:38:43.024117 2828 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 8 01:38:43.042716 kubelet[2828]: I1108 01:38:43.042626 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:43.045484 kubelet[2828]: I1108 01:38:43.045431 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:43.047214 kubelet[2828]: E1108 01:38:43.047124 2828 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:43.048067 kubelet[2828]: I1108 01:38:43.048026 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:43.049832 kubelet[2828]: E1108 01:38:43.049723 2828 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-3863d1ecf5\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:43.051848 kubelet[2828]: E1108 01:38:43.051753 2828 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:44.050900 kubelet[2828]: I1108 01:38:44.050847 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:44.051830 kubelet[2828]: I1108 01:38:44.051050 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:44.068741 kubelet[2828]: W1108 01:38:44.068685 2828 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:44.068920 kubelet[2828]: W1108 01:38:44.068681 2828 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:44.779074 kubelet[2828]: I1108 01:38:44.779022 2828 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:44.785575 kubelet[2828]: W1108 01:38:44.785521 2828 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:45.125400 systemd[1]: Reloading requested from client PID 3142 ('systemctl') (unit session-11.scope)... Nov 8 01:38:45.125433 systemd[1]: Reloading... Nov 8 01:38:45.185234 zram_generator::config[3181]: No configuration found. Nov 8 01:38:45.287319 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 8 01:38:45.353063 systemd[1]: Reloading finished in 226 ms. Nov 8 01:38:45.378277 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Nov 8 01:38:45.398418 systemd[1]: kubelet.service: Deactivated successfully. Nov 8 01:38:45.398568 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:45.418440 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 8 01:38:45.661530 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 8 01:38:45.665676 (kubelet)[3256]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 8 01:38:45.685515 kubelet[3256]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 8 01:38:45.685515 kubelet[3256]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 8 01:38:45.685515 kubelet[3256]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 8 01:38:45.685761 kubelet[3256]: I1108 01:38:45.685524 3256 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 8 01:38:45.689557 kubelet[3256]: I1108 01:38:45.689517 3256 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 8 01:38:45.689557 kubelet[3256]: I1108 01:38:45.689529 3256 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 8 01:38:45.689694 kubelet[3256]: I1108 01:38:45.689660 3256 server.go:954] "Client rotation is on, will bootstrap in background" Nov 8 01:38:45.690394 kubelet[3256]: I1108 01:38:45.690357 3256 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 8 01:38:45.691585 kubelet[3256]: I1108 01:38:45.691576 3256 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 8 01:38:45.693124 kubelet[3256]: E1108 01:38:45.693107 3256 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Nov 8 01:38:45.693164 kubelet[3256]: I1108 01:38:45.693125 3256 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Nov 8 01:38:45.700782 kubelet[3256]: I1108 01:38:45.700733 3256 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 8 01:38:45.701057 kubelet[3256]: I1108 01:38:45.701010 3256 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 8 01:38:45.701163 kubelet[3256]: I1108 01:38:45.701029 3256 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.6-n-3863d1ecf5","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Nov 8 01:38:45.701163 kubelet[3256]: I1108 01:38:45.701140 3256 topology_manager.go:138] "Creating topology manager with none policy" Nov 8 01:38:45.701163 kubelet[3256]: I1108 01:38:45.701154 3256 container_manager_linux.go:304] "Creating device plugin manager" Nov 8 01:38:45.701288 kubelet[3256]: I1108 01:38:45.701184 3256 state_mem.go:36] "Initialized new in-memory state store" Nov 8 01:38:45.701288 kubelet[3256]: I1108 01:38:45.701286 3256 kubelet.go:446] "Attempting to sync node with API server" Nov 8 01:38:45.701321 kubelet[3256]: I1108 01:38:45.701298 3256 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 8 01:38:45.701321 kubelet[3256]: I1108 01:38:45.701308 3256 kubelet.go:352] "Adding apiserver pod source" Nov 8 01:38:45.701321 kubelet[3256]: I1108 01:38:45.701316 3256 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 8 01:38:45.701790 kubelet[3256]: I1108 01:38:45.701759 3256 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Nov 8 01:38:45.702031 kubelet[3256]: I1108 01:38:45.702024 3256 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 8 01:38:45.702319 kubelet[3256]: I1108 01:38:45.702313 3256 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 8 01:38:45.702342 kubelet[3256]: I1108 01:38:45.702328 3256 server.go:1287] "Started kubelet" Nov 8 01:38:45.702431 kubelet[3256]: I1108 01:38:45.702408 3256 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 8 01:38:45.702594 kubelet[3256]: I1108 01:38:45.702507 3256 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 8 01:38:45.702762 kubelet[3256]: I1108 01:38:45.702747 3256 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 8 01:38:45.703520 kubelet[3256]: I1108 01:38:45.703511 3256 server.go:479] "Adding debug handlers to kubelet server" Nov 8 01:38:45.704481 kubelet[3256]: E1108 01:38:45.703949 3256 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 8 01:38:45.704481 kubelet[3256]: I1108 01:38:45.704101 3256 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 8 01:38:45.704481 kubelet[3256]: I1108 01:38:45.704370 3256 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 8 01:38:45.704481 kubelet[3256]: E1108 01:38:45.704459 3256 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.6-n-3863d1ecf5\" not found" Nov 8 01:38:45.704481 kubelet[3256]: I1108 01:38:45.704464 3256 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 8 01:38:45.704702 kubelet[3256]: I1108 01:38:45.704489 3256 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 8 01:38:45.705358 kubelet[3256]: I1108 01:38:45.705338 3256 reconciler.go:26] "Reconciler: start to sync state" Nov 8 01:38:45.705475 kubelet[3256]: I1108 01:38:45.705463 3256 factory.go:221] Registration of the systemd container factory successfully Nov 8 01:38:45.705535 kubelet[3256]: I1108 01:38:45.705524 3256 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 8 01:38:45.706132 kubelet[3256]: I1108 01:38:45.706122 3256 factory.go:221] Registration of the containerd container factory successfully Nov 8 01:38:45.709948 kubelet[3256]: I1108 01:38:45.709921 3256 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 8 01:38:45.710589 kubelet[3256]: I1108 01:38:45.710571 3256 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 8 01:38:45.710649 kubelet[3256]: I1108 01:38:45.710596 3256 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 8 01:38:45.710649 kubelet[3256]: I1108 01:38:45.710616 3256 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 8 01:38:45.710649 kubelet[3256]: I1108 01:38:45.710624 3256 kubelet.go:2382] "Starting kubelet main sync loop" Nov 8 01:38:45.710754 kubelet[3256]: E1108 01:38:45.710667 3256 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 8 01:38:45.726837 kubelet[3256]: I1108 01:38:45.726787 3256 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 8 01:38:45.726837 kubelet[3256]: I1108 01:38:45.726797 3256 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 8 01:38:45.726837 kubelet[3256]: I1108 01:38:45.726808 3256 state_mem.go:36] "Initialized new in-memory state store" Nov 8 01:38:45.726948 kubelet[3256]: I1108 01:38:45.726902 3256 state_mem.go:88] "Updated default CPUSet" cpuSet="" Nov 8 01:38:45.726948 kubelet[3256]: I1108 01:38:45.726909 3256 state_mem.go:96] "Updated CPUSet assignments" assignments={} Nov 8 01:38:45.726948 kubelet[3256]: I1108 01:38:45.726921 3256 policy_none.go:49] "None policy: Start" Nov 8 01:38:45.726948 kubelet[3256]: I1108 01:38:45.726927 3256 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 8 01:38:45.726948 kubelet[3256]: I1108 01:38:45.726932 3256 state_mem.go:35] "Initializing new in-memory state store" Nov 8 01:38:45.727033 kubelet[3256]: I1108 01:38:45.726989 3256 state_mem.go:75] "Updated machine memory state" Nov 8 01:38:45.727637 kubelet[3256]: I1108 01:38:45.727593 3256 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 8 01:38:45.727731 kubelet[3256]: I1108 01:38:45.727687 3256 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 8 01:38:45.727731 kubelet[3256]: I1108 01:38:45.727693 3256 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 8 01:38:45.727786 kubelet[3256]: I1108 01:38:45.727781 3256 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 8 01:38:45.728112 kubelet[3256]: E1108 01:38:45.728103 3256 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 8 01:38:45.812400 kubelet[3256]: I1108 01:38:45.812283 3256 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.812633 kubelet[3256]: I1108 01:38:45.812440 3256 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.812633 kubelet[3256]: I1108 01:38:45.812564 3256 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.821215 kubelet[3256]: W1108 01:38:45.821128 3256 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:45.821215 kubelet[3256]: W1108 01:38:45.821209 3256 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:45.821607 kubelet[3256]: W1108 01:38:45.821253 3256 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:45.821607 kubelet[3256]: E1108 01:38:45.821321 3256 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.821607 kubelet[3256]: E1108 01:38:45.821410 3256 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-3863d1ecf5\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.821607 kubelet[3256]: E1108 01:38:45.821332 3256 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" already exists" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.835219 kubelet[3256]: I1108 01:38:45.835123 3256 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.844531 kubelet[3256]: I1108 01:38:45.844477 3256 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.844721 kubelet[3256]: I1108 01:38:45.844621 3256 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.906966 kubelet[3256]: I1108 01:38:45.906879 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907249 kubelet[3256]: I1108 01:38:45.907025 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7922af8ce0cb5d65be29beff962b809e-ca-certs\") pod \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" (UID: \"7922af8ce0cb5d65be29beff962b809e\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907249 kubelet[3256]: I1108 01:38:45.907109 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7922af8ce0cb5d65be29beff962b809e-k8s-certs\") pod \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" (UID: \"7922af8ce0cb5d65be29beff962b809e\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907249 kubelet[3256]: I1108 01:38:45.907194 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-ca-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907604 kubelet[3256]: I1108 01:38:45.907252 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907604 kubelet[3256]: I1108 01:38:45.907339 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/67df4e4b57fbbb4d0e729ac8bd6036d6-kubeconfig\") pod \"kube-scheduler-ci-4081.3.6-n-3863d1ecf5\" (UID: \"67df4e4b57fbbb4d0e729ac8bd6036d6\") " pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907604 kubelet[3256]: I1108 01:38:45.907401 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7922af8ce0cb5d65be29beff962b809e-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" (UID: \"7922af8ce0cb5d65be29beff962b809e\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907604 kubelet[3256]: I1108 01:38:45.907449 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:45.907604 kubelet[3256]: I1108 01:38:45.907498 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/332b2243d2a6139435e1aec31ab74d79-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.6-n-3863d1ecf5\" (UID: \"332b2243d2a6139435e1aec31ab74d79\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:46.702322 kubelet[3256]: I1108 01:38:46.702191 3256 apiserver.go:52] "Watching apiserver" Nov 8 01:38:46.715358 kubelet[3256]: I1108 01:38:46.715330 3256 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:46.715433 kubelet[3256]: I1108 01:38:46.715416 3256 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:46.717643 kubelet[3256]: W1108 01:38:46.717629 3256 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:46.717704 kubelet[3256]: E1108 01:38:46.717657 3256 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-3863d1ecf5\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:46.717753 kubelet[3256]: W1108 01:38:46.717727 3256 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Nov 8 01:38:46.717784 kubelet[3256]: E1108 01:38:46.717752 3256 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-3863d1ecf5\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" Nov 8 01:38:46.728237 kubelet[3256]: I1108 01:38:46.728205 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.6-n-3863d1ecf5" podStartSLOduration=2.728193999 podStartE2EDuration="2.728193999s" podCreationTimestamp="2025-11-08 01:38:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-08 01:38:46.7281871 +0000 UTC m=+1.060367846" watchObservedRunningTime="2025-11-08 01:38:46.728193999 +0000 UTC m=+1.060374740" Nov 8 01:38:46.736085 kubelet[3256]: I1108 01:38:46.736055 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-3863d1ecf5" podStartSLOduration=2.7360456490000002 podStartE2EDuration="2.736045649s" podCreationTimestamp="2025-11-08 01:38:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-08 01:38:46.735964786 +0000 UTC m=+1.068145531" watchObservedRunningTime="2025-11-08 01:38:46.736045649 +0000 UTC m=+1.068226391" Nov 8 01:38:46.736187 kubelet[3256]: I1108 01:38:46.736097 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.6-n-3863d1ecf5" podStartSLOduration=2.736094227 podStartE2EDuration="2.736094227s" podCreationTimestamp="2025-11-08 01:38:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-08 01:38:46.732220996 +0000 UTC m=+1.064401742" watchObservedRunningTime="2025-11-08 01:38:46.736094227 +0000 UTC m=+1.068274968" Nov 8 01:38:46.806020 kubelet[3256]: I1108 01:38:46.805922 3256 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 8 01:38:50.179509 kubelet[3256]: I1108 01:38:50.179398 3256 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Nov 8 01:38:50.180444 containerd[1926]: time="2025-11-08T01:38:50.180117644Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Nov 8 01:38:50.181083 kubelet[3256]: I1108 01:38:50.180548 3256 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Nov 8 01:38:51.046949 kubelet[3256]: I1108 01:38:51.046833 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da200673-cc1e-42c5-9173-31f7fb657552-lib-modules\") pod \"kube-proxy-4lwvd\" (UID: \"da200673-cc1e-42c5-9173-31f7fb657552\") " pod="kube-system/kube-proxy-4lwvd" Nov 8 01:38:51.046949 kubelet[3256]: I1108 01:38:51.046937 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/da200673-cc1e-42c5-9173-31f7fb657552-xtables-lock\") pod \"kube-proxy-4lwvd\" (UID: \"da200673-cc1e-42c5-9173-31f7fb657552\") " pod="kube-system/kube-proxy-4lwvd" Nov 8 01:38:51.047308 kubelet[3256]: I1108 01:38:51.046991 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/da200673-cc1e-42c5-9173-31f7fb657552-kube-proxy\") pod \"kube-proxy-4lwvd\" (UID: \"da200673-cc1e-42c5-9173-31f7fb657552\") " pod="kube-system/kube-proxy-4lwvd" Nov 8 01:38:51.047308 kubelet[3256]: I1108 01:38:51.047106 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwmf\" (UniqueName: \"kubernetes.io/projected/da200673-cc1e-42c5-9173-31f7fb657552-kube-api-access-6xwmf\") pod \"kube-proxy-4lwvd\" (UID: \"da200673-cc1e-42c5-9173-31f7fb657552\") " pod="kube-system/kube-proxy-4lwvd" Nov 8 01:38:51.244602 containerd[1926]: time="2025-11-08T01:38:51.244433713Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4lwvd,Uid:da200673-cc1e-42c5-9173-31f7fb657552,Namespace:kube-system,Attempt:0,}" Nov 8 01:38:51.260362 containerd[1926]: time="2025-11-08T01:38:51.260300052Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:38:51.260362 containerd[1926]: time="2025-11-08T01:38:51.260329632Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:38:51.260362 containerd[1926]: time="2025-11-08T01:38:51.260336747Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:51.260468 containerd[1926]: time="2025-11-08T01:38:51.260384754Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:51.298268 containerd[1926]: time="2025-11-08T01:38:51.298172071Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4lwvd,Uid:da200673-cc1e-42c5-9173-31f7fb657552,Namespace:kube-system,Attempt:0,} returns sandbox id \"6763a5857b97b1fd51c82f9250ba799fa9599a3405d17530228be34fd61f1590\"" Nov 8 01:38:51.299646 containerd[1926]: time="2025-11-08T01:38:51.299628234Z" level=info msg="CreateContainer within sandbox \"6763a5857b97b1fd51c82f9250ba799fa9599a3405d17530228be34fd61f1590\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Nov 8 01:38:51.304982 containerd[1926]: time="2025-11-08T01:38:51.304934470Z" level=info msg="CreateContainer within sandbox \"6763a5857b97b1fd51c82f9250ba799fa9599a3405d17530228be34fd61f1590\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"087a07931c039a325f81e0fef684f4a0f3b534c2582a3d59645f92bd9a543572\"" Nov 8 01:38:51.305215 containerd[1926]: time="2025-11-08T01:38:51.305198876Z" level=info msg="StartContainer for \"087a07931c039a325f81e0fef684f4a0f3b534c2582a3d59645f92bd9a543572\"" Nov 8 01:38:51.349105 kubelet[3256]: I1108 01:38:51.349076 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/9c8aa86a-80f9-4a09-9646-eaf95e771867-var-lib-calico\") pod \"tigera-operator-7dcd859c48-9jtcr\" (UID: \"9c8aa86a-80f9-4a09-9646-eaf95e771867\") " pod="tigera-operator/tigera-operator-7dcd859c48-9jtcr" Nov 8 01:38:51.349416 kubelet[3256]: I1108 01:38:51.349112 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg7zt\" (UniqueName: \"kubernetes.io/projected/9c8aa86a-80f9-4a09-9646-eaf95e771867-kube-api-access-lg7zt\") pod \"tigera-operator-7dcd859c48-9jtcr\" (UID: \"9c8aa86a-80f9-4a09-9646-eaf95e771867\") " pod="tigera-operator/tigera-operator-7dcd859c48-9jtcr" Nov 8 01:38:51.356260 containerd[1926]: time="2025-11-08T01:38:51.356234341Z" level=info msg="StartContainer for \"087a07931c039a325f81e0fef684f4a0f3b534c2582a3d59645f92bd9a543572\" returns successfully" Nov 8 01:38:51.629057 containerd[1926]: time="2025-11-08T01:38:51.628815387Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-9jtcr,Uid:9c8aa86a-80f9-4a09-9646-eaf95e771867,Namespace:tigera-operator,Attempt:0,}" Nov 8 01:38:51.639331 containerd[1926]: time="2025-11-08T01:38:51.639091345Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:38:51.639331 containerd[1926]: time="2025-11-08T01:38:51.639316728Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:38:51.639331 containerd[1926]: time="2025-11-08T01:38:51.639328720Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:51.639445 containerd[1926]: time="2025-11-08T01:38:51.639379842Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:38:51.687254 containerd[1926]: time="2025-11-08T01:38:51.687232957Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-9jtcr,Uid:9c8aa86a-80f9-4a09-9646-eaf95e771867,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"2bb61b0b558eacadda3016996331964733e2b66a7b7e90112f676b999e4322da\"" Nov 8 01:38:51.688028 containerd[1926]: time="2025-11-08T01:38:51.688014250Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Nov 8 01:38:52.881815 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount86781685.mount: Deactivated successfully. Nov 8 01:38:53.450601 containerd[1926]: time="2025-11-08T01:38:53.450550844Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:53.450810 containerd[1926]: time="2025-11-08T01:38:53.450742843Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=25061691" Nov 8 01:38:53.451182 containerd[1926]: time="2025-11-08T01:38:53.451126350Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:53.452296 containerd[1926]: time="2025-11-08T01:38:53.452250412Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:38:53.452739 containerd[1926]: time="2025-11-08T01:38:53.452697312Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 1.764663016s" Nov 8 01:38:53.452739 containerd[1926]: time="2025-11-08T01:38:53.452712762Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Nov 8 01:38:53.453716 containerd[1926]: time="2025-11-08T01:38:53.453703315Z" level=info msg="CreateContainer within sandbox \"2bb61b0b558eacadda3016996331964733e2b66a7b7e90112f676b999e4322da\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Nov 8 01:38:53.457259 containerd[1926]: time="2025-11-08T01:38:53.457244763Z" level=info msg="CreateContainer within sandbox \"2bb61b0b558eacadda3016996331964733e2b66a7b7e90112f676b999e4322da\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"c1ad835559c4da02afb349fd8a42af1e41be4df2066caae3b873c2a341a399ea\"" Nov 8 01:38:53.457519 containerd[1926]: time="2025-11-08T01:38:53.457506231Z" level=info msg="StartContainer for \"c1ad835559c4da02afb349fd8a42af1e41be4df2066caae3b873c2a341a399ea\"" Nov 8 01:38:53.487624 containerd[1926]: time="2025-11-08T01:38:53.487603230Z" level=info msg="StartContainer for \"c1ad835559c4da02afb349fd8a42af1e41be4df2066caae3b873c2a341a399ea\" returns successfully" Nov 8 01:38:53.625456 kubelet[3256]: I1108 01:38:53.625425 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-4lwvd" podStartSLOduration=3.6254158260000002 podStartE2EDuration="3.625415826s" podCreationTimestamp="2025-11-08 01:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-08 01:38:51.748743407 +0000 UTC m=+6.080924219" watchObservedRunningTime="2025-11-08 01:38:53.625415826 +0000 UTC m=+7.957596572" Nov 8 01:38:53.747779 kubelet[3256]: I1108 01:38:53.747706 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-9jtcr" podStartSLOduration=0.982314943 podStartE2EDuration="2.747696712s" podCreationTimestamp="2025-11-08 01:38:51 +0000 UTC" firstStartedPulling="2025-11-08 01:38:51.687784653 +0000 UTC m=+6.019965398" lastFinishedPulling="2025-11-08 01:38:53.453166427 +0000 UTC m=+7.785347167" observedRunningTime="2025-11-08 01:38:53.74763273 +0000 UTC m=+8.079813475" watchObservedRunningTime="2025-11-08 01:38:53.747696712 +0000 UTC m=+8.079877454" Nov 8 01:38:57.131541 update_engine[1914]: I20251108 01:38:57.131371 1914 update_attempter.cc:509] Updating boot flags... Nov 8 01:38:57.187152 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 35 scanned by (udev-worker) (3738) Nov 8 01:38:57.213150 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 35 scanned by (udev-worker) (3739) Nov 8 01:38:57.826966 sudo[2244]: pam_unix(sudo:session): session closed for user root Nov 8 01:38:57.828044 sshd[2238]: pam_unix(sshd:session): session closed for user core Nov 8 01:38:57.829639 systemd[1]: sshd@8-139.178.94.189:22-139.178.68.195:58078.service: Deactivated successfully. Nov 8 01:38:57.831067 systemd-logind[1909]: Session 11 logged out. Waiting for processes to exit. Nov 8 01:38:57.831157 systemd[1]: session-11.scope: Deactivated successfully. Nov 8 01:38:57.831650 systemd-logind[1909]: Removed session 11. Nov 8 01:39:03.431359 kubelet[3256]: I1108 01:39:03.431286 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea5190a8-8edf-4c79-a1b2-c7ce403371aa-tigera-ca-bundle\") pod \"calico-typha-74c857bf99-pw2mn\" (UID: \"ea5190a8-8edf-4c79-a1b2-c7ce403371aa\") " pod="calico-system/calico-typha-74c857bf99-pw2mn" Nov 8 01:39:03.431359 kubelet[3256]: I1108 01:39:03.431354 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/ea5190a8-8edf-4c79-a1b2-c7ce403371aa-typha-certs\") pod \"calico-typha-74c857bf99-pw2mn\" (UID: \"ea5190a8-8edf-4c79-a1b2-c7ce403371aa\") " pod="calico-system/calico-typha-74c857bf99-pw2mn" Nov 8 01:39:03.431788 kubelet[3256]: I1108 01:39:03.431389 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d54rg\" (UniqueName: \"kubernetes.io/projected/ea5190a8-8edf-4c79-a1b2-c7ce403371aa-kube-api-access-d54rg\") pod \"calico-typha-74c857bf99-pw2mn\" (UID: \"ea5190a8-8edf-4c79-a1b2-c7ce403371aa\") " pod="calico-system/calico-typha-74c857bf99-pw2mn" Nov 8 01:39:03.633461 kubelet[3256]: I1108 01:39:03.633330 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-xtables-lock\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.633461 kubelet[3256]: I1108 01:39:03.633434 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-cni-bin-dir\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.633838 kubelet[3256]: I1108 01:39:03.633501 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5k2j\" (UniqueName: \"kubernetes.io/projected/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-kube-api-access-n5k2j\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.633838 kubelet[3256]: I1108 01:39:03.633590 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-policysync\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.633838 kubelet[3256]: I1108 01:39:03.633644 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-lib-modules\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.633838 kubelet[3256]: I1108 01:39:03.633750 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-tigera-ca-bundle\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.634285 kubelet[3256]: I1108 01:39:03.633857 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-var-run-calico\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.634285 kubelet[3256]: I1108 01:39:03.633938 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-cni-net-dir\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.634285 kubelet[3256]: I1108 01:39:03.633997 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-flexvol-driver-host\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.634285 kubelet[3256]: I1108 01:39:03.634064 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-node-certs\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.634285 kubelet[3256]: I1108 01:39:03.634170 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-cni-log-dir\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.634779 kubelet[3256]: I1108 01:39:03.634221 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60-var-lib-calico\") pod \"calico-node-6fmld\" (UID: \"f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60\") " pod="calico-system/calico-node-6fmld" Nov 8 01:39:03.698393 containerd[1926]: time="2025-11-08T01:39:03.698123727Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-74c857bf99-pw2mn,Uid:ea5190a8-8edf-4c79-a1b2-c7ce403371aa,Namespace:calico-system,Attempt:0,}" Nov 8 01:39:03.736148 kubelet[3256]: E1108 01:39:03.736095 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.736148 kubelet[3256]: W1108 01:39:03.736111 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.736148 kubelet[3256]: E1108 01:39:03.736136 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.737074 kubelet[3256]: E1108 01:39:03.737061 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.737074 kubelet[3256]: W1108 01:39:03.737070 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.737172 kubelet[3256]: E1108 01:39:03.737078 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.741400 kubelet[3256]: E1108 01:39:03.741386 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.741400 kubelet[3256]: W1108 01:39:03.741395 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.741511 kubelet[3256]: E1108 01:39:03.741404 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.758436 kubelet[3256]: E1108 01:39:03.758391 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:03.821154 kubelet[3256]: E1108 01:39:03.821087 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.821154 kubelet[3256]: W1108 01:39:03.821112 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.821154 kubelet[3256]: E1108 01:39:03.821133 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.821492 kubelet[3256]: E1108 01:39:03.821438 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.821492 kubelet[3256]: W1108 01:39:03.821456 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.821492 kubelet[3256]: E1108 01:39:03.821472 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.821811 kubelet[3256]: E1108 01:39:03.821757 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.821811 kubelet[3256]: W1108 01:39:03.821775 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.821811 kubelet[3256]: E1108 01:39:03.821791 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.822102 kubelet[3256]: E1108 01:39:03.822086 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.822102 kubelet[3256]: W1108 01:39:03.822100 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.822237 kubelet[3256]: E1108 01:39:03.822115 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.822517 kubelet[3256]: E1108 01:39:03.822462 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.822517 kubelet[3256]: W1108 01:39:03.822483 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.822517 kubelet[3256]: E1108 01:39:03.822502 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.822888 kubelet[3256]: E1108 01:39:03.822837 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.822888 kubelet[3256]: W1108 01:39:03.822852 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.822888 kubelet[3256]: E1108 01:39:03.822867 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.823134 kubelet[3256]: E1108 01:39:03.823115 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.823134 kubelet[3256]: W1108 01:39:03.823134 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.823298 kubelet[3256]: E1108 01:39:03.823171 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.823409 kubelet[3256]: E1108 01:39:03.823394 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.823474 kubelet[3256]: W1108 01:39:03.823408 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.823474 kubelet[3256]: E1108 01:39:03.823424 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.823663 kubelet[3256]: E1108 01:39:03.823648 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.823722 kubelet[3256]: W1108 01:39:03.823662 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.823722 kubelet[3256]: E1108 01:39:03.823676 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.823891 kubelet[3256]: E1108 01:39:03.823877 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.823956 kubelet[3256]: W1108 01:39:03.823891 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.823956 kubelet[3256]: E1108 01:39:03.823905 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.824116 kubelet[3256]: E1108 01:39:03.824102 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.824189 kubelet[3256]: W1108 01:39:03.824116 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.824189 kubelet[3256]: E1108 01:39:03.824129 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.824435 kubelet[3256]: E1108 01:39:03.824420 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.824501 kubelet[3256]: W1108 01:39:03.824435 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.824501 kubelet[3256]: E1108 01:39:03.824448 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.824757 kubelet[3256]: E1108 01:39:03.824741 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.824812 kubelet[3256]: W1108 01:39:03.824757 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.824812 kubelet[3256]: E1108 01:39:03.824771 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.824990 kubelet[3256]: E1108 01:39:03.824976 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.825055 kubelet[3256]: W1108 01:39:03.824990 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.825055 kubelet[3256]: E1108 01:39:03.825004 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.825230 kubelet[3256]: E1108 01:39:03.825216 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.825303 kubelet[3256]: W1108 01:39:03.825230 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.825303 kubelet[3256]: E1108 01:39:03.825244 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.825528 kubelet[3256]: E1108 01:39:03.825514 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.825596 kubelet[3256]: W1108 01:39:03.825528 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.825596 kubelet[3256]: E1108 01:39:03.825541 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.825879 kubelet[3256]: E1108 01:39:03.825839 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.825879 kubelet[3256]: W1108 01:39:03.825854 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.825879 kubelet[3256]: E1108 01:39:03.825870 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.826084 kubelet[3256]: E1108 01:39:03.826071 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.826164 kubelet[3256]: W1108 01:39:03.826084 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.826164 kubelet[3256]: E1108 01:39:03.826097 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.826377 kubelet[3256]: E1108 01:39:03.826339 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.826377 kubelet[3256]: W1108 01:39:03.826353 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.826377 kubelet[3256]: E1108 01:39:03.826368 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.826610 kubelet[3256]: E1108 01:39:03.826573 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.826610 kubelet[3256]: W1108 01:39:03.826587 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.826610 kubelet[3256]: E1108 01:39:03.826599 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.835918 kubelet[3256]: E1108 01:39:03.835882 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.835918 kubelet[3256]: W1108 01:39:03.835892 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.835918 kubelet[3256]: E1108 01:39:03.835901 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.836020 kubelet[3256]: I1108 01:39:03.835921 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lrzl\" (UniqueName: \"kubernetes.io/projected/ff4b01f8-4bf6-4050-afe4-e2ac586eece3-kube-api-access-9lrzl\") pod \"csi-node-driver-dmqbg\" (UID: \"ff4b01f8-4bf6-4050-afe4-e2ac586eece3\") " pod="calico-system/csi-node-driver-dmqbg" Nov 8 01:39:03.836072 kubelet[3256]: E1108 01:39:03.836063 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.836104 kubelet[3256]: W1108 01:39:03.836072 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.836104 kubelet[3256]: E1108 01:39:03.836082 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.836104 kubelet[3256]: I1108 01:39:03.836097 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/ff4b01f8-4bf6-4050-afe4-e2ac586eece3-varrun\") pod \"csi-node-driver-dmqbg\" (UID: \"ff4b01f8-4bf6-4050-afe4-e2ac586eece3\") " pod="calico-system/csi-node-driver-dmqbg" Nov 8 01:39:03.836322 kubelet[3256]: E1108 01:39:03.836285 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.836322 kubelet[3256]: W1108 01:39:03.836296 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.836322 kubelet[3256]: E1108 01:39:03.836308 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.836423 kubelet[3256]: I1108 01:39:03.836324 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff4b01f8-4bf6-4050-afe4-e2ac586eece3-kubelet-dir\") pod \"csi-node-driver-dmqbg\" (UID: \"ff4b01f8-4bf6-4050-afe4-e2ac586eece3\") " pod="calico-system/csi-node-driver-dmqbg" Nov 8 01:39:03.836555 kubelet[3256]: E1108 01:39:03.836510 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.836555 kubelet[3256]: W1108 01:39:03.836521 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.836555 kubelet[3256]: E1108 01:39:03.836533 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.836555 kubelet[3256]: I1108 01:39:03.836548 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ff4b01f8-4bf6-4050-afe4-e2ac586eece3-socket-dir\") pod \"csi-node-driver-dmqbg\" (UID: \"ff4b01f8-4bf6-4050-afe4-e2ac586eece3\") " pod="calico-system/csi-node-driver-dmqbg" Nov 8 01:39:03.836756 kubelet[3256]: E1108 01:39:03.836712 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.836756 kubelet[3256]: W1108 01:39:03.836721 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.836756 kubelet[3256]: E1108 01:39:03.836731 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.836756 kubelet[3256]: I1108 01:39:03.836744 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ff4b01f8-4bf6-4050-afe4-e2ac586eece3-registration-dir\") pod \"csi-node-driver-dmqbg\" (UID: \"ff4b01f8-4bf6-4050-afe4-e2ac586eece3\") " pod="calico-system/csi-node-driver-dmqbg" Nov 8 01:39:03.836884 kubelet[3256]: E1108 01:39:03.836875 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.836884 kubelet[3256]: W1108 01:39:03.836883 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.836946 kubelet[3256]: E1108 01:39:03.836892 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.837014 kubelet[3256]: E1108 01:39:03.837007 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.837040 kubelet[3256]: W1108 01:39:03.837014 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.837067 kubelet[3256]: E1108 01:39:03.837033 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.837194 kubelet[3256]: E1108 01:39:03.837187 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.837230 kubelet[3256]: W1108 01:39:03.837194 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.837230 kubelet[3256]: E1108 01:39:03.837212 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.837371 kubelet[3256]: E1108 01:39:03.837363 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.837400 kubelet[3256]: W1108 01:39:03.837370 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.837400 kubelet[3256]: E1108 01:39:03.837380 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.837545 kubelet[3256]: E1108 01:39:03.837537 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.837578 kubelet[3256]: W1108 01:39:03.837544 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.837578 kubelet[3256]: E1108 01:39:03.837553 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.837685 kubelet[3256]: E1108 01:39:03.837677 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.837715 kubelet[3256]: W1108 01:39:03.837685 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.837715 kubelet[3256]: E1108 01:39:03.837693 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.837843 kubelet[3256]: E1108 01:39:03.837836 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.837877 kubelet[3256]: W1108 01:39:03.837843 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.837877 kubelet[3256]: E1108 01:39:03.837850 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.837974 kubelet[3256]: E1108 01:39:03.837967 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.838002 kubelet[3256]: W1108 01:39:03.837974 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.838002 kubelet[3256]: E1108 01:39:03.837982 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.838102 kubelet[3256]: E1108 01:39:03.838095 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.838139 kubelet[3256]: W1108 01:39:03.838102 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.838139 kubelet[3256]: E1108 01:39:03.838110 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.838255 kubelet[3256]: E1108 01:39:03.838247 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.838255 kubelet[3256]: W1108 01:39:03.838255 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.838318 kubelet[3256]: E1108 01:39:03.838262 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.877162 containerd[1926]: time="2025-11-08T01:39:03.873487916Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6fmld,Uid:f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60,Namespace:calico-system,Attempt:0,}" Nov 8 01:39:03.937743 kubelet[3256]: E1108 01:39:03.937690 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.937743 kubelet[3256]: W1108 01:39:03.937702 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.937743 kubelet[3256]: E1108 01:39:03.937715 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940013 kubelet[3256]: E1108 01:39:03.937891 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940013 kubelet[3256]: W1108 01:39:03.937898 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940013 kubelet[3256]: E1108 01:39:03.937907 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940013 kubelet[3256]: E1108 01:39:03.938015 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940013 kubelet[3256]: W1108 01:39:03.938020 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940013 kubelet[3256]: E1108 01:39:03.938026 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940013 kubelet[3256]: E1108 01:39:03.938115 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940013 kubelet[3256]: W1108 01:39:03.938119 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940013 kubelet[3256]: E1108 01:39:03.938124 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940013 kubelet[3256]: E1108 01:39:03.938211 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940195 kubelet[3256]: W1108 01:39:03.938218 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940195 kubelet[3256]: E1108 01:39:03.938226 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940195 kubelet[3256]: E1108 01:39:03.938354 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940195 kubelet[3256]: W1108 01:39:03.938360 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940195 kubelet[3256]: E1108 01:39:03.938367 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940195 kubelet[3256]: E1108 01:39:03.938441 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940195 kubelet[3256]: W1108 01:39:03.938447 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940195 kubelet[3256]: E1108 01:39:03.938454 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940195 kubelet[3256]: E1108 01:39:03.938544 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940195 kubelet[3256]: W1108 01:39:03.938552 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940351 kubelet[3256]: E1108 01:39:03.938559 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940351 kubelet[3256]: E1108 01:39:03.938636 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940351 kubelet[3256]: W1108 01:39:03.938641 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940351 kubelet[3256]: E1108 01:39:03.938646 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940351 kubelet[3256]: E1108 01:39:03.938755 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940351 kubelet[3256]: W1108 01:39:03.938759 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940351 kubelet[3256]: E1108 01:39:03.938764 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940351 kubelet[3256]: E1108 01:39:03.938885 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940351 kubelet[3256]: W1108 01:39:03.938890 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940351 kubelet[3256]: E1108 01:39:03.938895 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940524 kubelet[3256]: E1108 01:39:03.939034 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940524 kubelet[3256]: W1108 01:39:03.939040 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940524 kubelet[3256]: E1108 01:39:03.939047 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940524 kubelet[3256]: E1108 01:39:03.939123 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940524 kubelet[3256]: W1108 01:39:03.939128 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940524 kubelet[3256]: E1108 01:39:03.939134 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940524 kubelet[3256]: E1108 01:39:03.939211 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940524 kubelet[3256]: W1108 01:39:03.939215 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940524 kubelet[3256]: E1108 01:39:03.939225 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940524 kubelet[3256]: E1108 01:39:03.939323 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940691 kubelet[3256]: W1108 01:39:03.939327 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940691 kubelet[3256]: E1108 01:39:03.939335 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940691 kubelet[3256]: E1108 01:39:03.939397 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940691 kubelet[3256]: W1108 01:39:03.939401 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940691 kubelet[3256]: E1108 01:39:03.939412 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940691 kubelet[3256]: E1108 01:39:03.939490 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940691 kubelet[3256]: W1108 01:39:03.939494 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940691 kubelet[3256]: E1108 01:39:03.939500 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940691 kubelet[3256]: E1108 01:39:03.939589 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940691 kubelet[3256]: W1108 01:39:03.939594 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940871 kubelet[3256]: E1108 01:39:03.939602 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940871 kubelet[3256]: E1108 01:39:03.939670 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940871 kubelet[3256]: W1108 01:39:03.939675 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940871 kubelet[3256]: E1108 01:39:03.939680 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940871 kubelet[3256]: E1108 01:39:03.939743 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940871 kubelet[3256]: W1108 01:39:03.939747 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940871 kubelet[3256]: E1108 01:39:03.939753 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.940871 kubelet[3256]: E1108 01:39:03.939815 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.940871 kubelet[3256]: W1108 01:39:03.939819 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.940871 kubelet[3256]: E1108 01:39:03.939824 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.941030 kubelet[3256]: E1108 01:39:03.939895 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.941030 kubelet[3256]: W1108 01:39:03.939899 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.941030 kubelet[3256]: E1108 01:39:03.939906 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.941030 kubelet[3256]: E1108 01:39:03.939976 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.941030 kubelet[3256]: W1108 01:39:03.939981 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.941030 kubelet[3256]: E1108 01:39:03.939987 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.941030 kubelet[3256]: E1108 01:39:03.940140 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.941030 kubelet[3256]: W1108 01:39:03.940151 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.941030 kubelet[3256]: E1108 01:39:03.940158 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.941030 kubelet[3256]: E1108 01:39:03.940291 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.941213 kubelet[3256]: W1108 01:39:03.940296 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.941213 kubelet[3256]: E1108 01:39:03.940301 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.944431 kubelet[3256]: E1108 01:39:03.944421 3256 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Nov 8 01:39:03.944431 kubelet[3256]: W1108 01:39:03.944429 3256 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Nov 8 01:39:03.945049 kubelet[3256]: E1108 01:39:03.944439 3256 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Nov 8 01:39:03.953039 containerd[1926]: time="2025-11-08T01:39:03.952959473Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:03.953039 containerd[1926]: time="2025-11-08T01:39:03.952990331Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:03.953039 containerd[1926]: time="2025-11-08T01:39:03.953015279Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:03.953268 containerd[1926]: time="2025-11-08T01:39:03.953249846Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:03.956343 containerd[1926]: time="2025-11-08T01:39:03.956294670Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:03.956529 containerd[1926]: time="2025-11-08T01:39:03.956511855Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:03.956553 containerd[1926]: time="2025-11-08T01:39:03.956526141Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:03.956586 containerd[1926]: time="2025-11-08T01:39:03.956574974Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:03.981338 containerd[1926]: time="2025-11-08T01:39:03.981315489Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6fmld,Uid:f55ef9a4-e0a6-4f0f-8fcb-a9bb919e1a60,Namespace:calico-system,Attempt:0,} returns sandbox id \"0e8517d4481369d52be9064628ad07d03821e659106ff00999716de429622f6d\"" Nov 8 01:39:03.982034 containerd[1926]: time="2025-11-08T01:39:03.982021120Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Nov 8 01:39:03.994114 containerd[1926]: time="2025-11-08T01:39:03.994090921Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-74c857bf99-pw2mn,Uid:ea5190a8-8edf-4c79-a1b2-c7ce403371aa,Namespace:calico-system,Attempt:0,} returns sandbox id \"3717003908c1aa5b6d78ec488e54d3f2cab8b5df5e79931235ee5096af69681a\"" Nov 8 01:39:05.639653 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1946348841.mount: Deactivated successfully. Nov 8 01:39:05.670890 containerd[1926]: time="2025-11-08T01:39:05.670841569Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:05.671109 containerd[1926]: time="2025-11-08T01:39:05.671078538Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=5941492" Nov 8 01:39:05.671432 containerd[1926]: time="2025-11-08T01:39:05.671392252Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:05.672392 containerd[1926]: time="2025-11-08T01:39:05.672352112Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:05.673137 containerd[1926]: time="2025-11-08T01:39:05.673095309Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.691055309s" Nov 8 01:39:05.673137 containerd[1926]: time="2025-11-08T01:39:05.673112635Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Nov 8 01:39:05.673558 containerd[1926]: time="2025-11-08T01:39:05.673547905Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Nov 8 01:39:05.674088 containerd[1926]: time="2025-11-08T01:39:05.674075879Z" level=info msg="CreateContainer within sandbox \"0e8517d4481369d52be9064628ad07d03821e659106ff00999716de429622f6d\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Nov 8 01:39:05.678521 containerd[1926]: time="2025-11-08T01:39:05.678479583Z" level=info msg="CreateContainer within sandbox \"0e8517d4481369d52be9064628ad07d03821e659106ff00999716de429622f6d\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"38a149eac9ef739cd0660243ed572e2895805f002e6d75acf2f5616936a66060\"" Nov 8 01:39:05.678731 containerd[1926]: time="2025-11-08T01:39:05.678716236Z" level=info msg="StartContainer for \"38a149eac9ef739cd0660243ed572e2895805f002e6d75acf2f5616936a66060\"" Nov 8 01:39:05.711355 kubelet[3256]: E1108 01:39:05.711284 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:05.716684 containerd[1926]: time="2025-11-08T01:39:05.716660722Z" level=info msg="StartContainer for \"38a149eac9ef739cd0660243ed572e2895805f002e6d75acf2f5616936a66060\" returns successfully" Nov 8 01:39:05.747089 containerd[1926]: time="2025-11-08T01:39:05.747066622Z" level=error msg="collecting metrics for 38a149eac9ef739cd0660243ed572e2895805f002e6d75acf2f5616936a66060" error="cgroups: cgroup deleted: unknown" Nov 8 01:39:06.633254 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-38a149eac9ef739cd0660243ed572e2895805f002e6d75acf2f5616936a66060-rootfs.mount: Deactivated successfully. Nov 8 01:39:06.700157 containerd[1926]: time="2025-11-08T01:39:06.700121272Z" level=info msg="shim disconnected" id=38a149eac9ef739cd0660243ed572e2895805f002e6d75acf2f5616936a66060 namespace=k8s.io Nov 8 01:39:06.700157 containerd[1926]: time="2025-11-08T01:39:06.700156106Z" level=warning msg="cleaning up after shim disconnected" id=38a149eac9ef739cd0660243ed572e2895805f002e6d75acf2f5616936a66060 namespace=k8s.io Nov 8 01:39:06.700157 containerd[1926]: time="2025-11-08T01:39:06.700162217Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 8 01:39:07.711850 kubelet[3256]: E1108 01:39:07.711796 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:08.116154 containerd[1926]: time="2025-11-08T01:39:08.116103294Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:08.116374 containerd[1926]: time="2025-11-08T01:39:08.116334053Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=33739890" Nov 8 01:39:08.116672 containerd[1926]: time="2025-11-08T01:39:08.116660370Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:08.117641 containerd[1926]: time="2025-11-08T01:39:08.117627831Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:08.118038 containerd[1926]: time="2025-11-08T01:39:08.118027225Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 2.444465495s" Nov 8 01:39:08.118069 containerd[1926]: time="2025-11-08T01:39:08.118038812Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Nov 8 01:39:08.118555 containerd[1926]: time="2025-11-08T01:39:08.118546635Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Nov 8 01:39:08.121557 containerd[1926]: time="2025-11-08T01:39:08.121541314Z" level=info msg="CreateContainer within sandbox \"3717003908c1aa5b6d78ec488e54d3f2cab8b5df5e79931235ee5096af69681a\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Nov 8 01:39:08.125417 containerd[1926]: time="2025-11-08T01:39:08.125375648Z" level=info msg="CreateContainer within sandbox \"3717003908c1aa5b6d78ec488e54d3f2cab8b5df5e79931235ee5096af69681a\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"3b64bb783cded4c2866a9079dd1fd4c274724f876e4ca4db11c487befa553d2f\"" Nov 8 01:39:08.125573 containerd[1926]: time="2025-11-08T01:39:08.125551071Z" level=info msg="StartContainer for \"3b64bb783cded4c2866a9079dd1fd4c274724f876e4ca4db11c487befa553d2f\"" Nov 8 01:39:08.175694 containerd[1926]: time="2025-11-08T01:39:08.175671808Z" level=info msg="StartContainer for \"3b64bb783cded4c2866a9079dd1fd4c274724f876e4ca4db11c487befa553d2f\" returns successfully" Nov 8 01:39:08.789001 kubelet[3256]: I1108 01:39:08.788853 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-74c857bf99-pw2mn" podStartSLOduration=1.664878759 podStartE2EDuration="5.788800101s" podCreationTimestamp="2025-11-08 01:39:03 +0000 UTC" firstStartedPulling="2025-11-08 01:39:03.994575639 +0000 UTC m=+18.326756386" lastFinishedPulling="2025-11-08 01:39:08.118496984 +0000 UTC m=+22.450677728" observedRunningTime="2025-11-08 01:39:08.788320955 +0000 UTC m=+23.120501776" watchObservedRunningTime="2025-11-08 01:39:08.788800101 +0000 UTC m=+23.120980940" Nov 8 01:39:09.711133 kubelet[3256]: E1108 01:39:09.711087 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:09.769048 kubelet[3256]: I1108 01:39:09.768972 3256 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 8 01:39:11.577792 containerd[1926]: time="2025-11-08T01:39:11.577740379Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:11.578004 containerd[1926]: time="2025-11-08T01:39:11.577970310Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70446859" Nov 8 01:39:11.578229 containerd[1926]: time="2025-11-08T01:39:11.578192725Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:11.579331 containerd[1926]: time="2025-11-08T01:39:11.579288134Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:11.579722 containerd[1926]: time="2025-11-08T01:39:11.579683697Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 3.461124766s" Nov 8 01:39:11.579722 containerd[1926]: time="2025-11-08T01:39:11.579697310Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Nov 8 01:39:11.580641 containerd[1926]: time="2025-11-08T01:39:11.580626351Z" level=info msg="CreateContainer within sandbox \"0e8517d4481369d52be9064628ad07d03821e659106ff00999716de429622f6d\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Nov 8 01:39:11.585198 containerd[1926]: time="2025-11-08T01:39:11.585168052Z" level=info msg="CreateContainer within sandbox \"0e8517d4481369d52be9064628ad07d03821e659106ff00999716de429622f6d\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"d692cb2b654e18a1fa4cd291557e3a065c309a087760dc92a27810c2eef9f723\"" Nov 8 01:39:11.585440 containerd[1926]: time="2025-11-08T01:39:11.585397405Z" level=info msg="StartContainer for \"d692cb2b654e18a1fa4cd291557e3a065c309a087760dc92a27810c2eef9f723\"" Nov 8 01:39:11.659906 containerd[1926]: time="2025-11-08T01:39:11.659843079Z" level=info msg="StartContainer for \"d692cb2b654e18a1fa4cd291557e3a065c309a087760dc92a27810c2eef9f723\" returns successfully" Nov 8 01:39:11.711712 kubelet[3256]: E1108 01:39:11.711642 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:12.272011 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d692cb2b654e18a1fa4cd291557e3a065c309a087760dc92a27810c2eef9f723-rootfs.mount: Deactivated successfully. Nov 8 01:39:12.288709 kubelet[3256]: I1108 01:39:12.288666 3256 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Nov 8 01:39:12.405612 kubelet[3256]: I1108 01:39:12.405476 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rplv5\" (UniqueName: \"kubernetes.io/projected/5f897823-41ac-4b66-8214-3b03dd8d6290-kube-api-access-rplv5\") pod \"goldmane-666569f655-8ld4r\" (UID: \"5f897823-41ac-4b66-8214-3b03dd8d6290\") " pod="calico-system/goldmane-666569f655-8ld4r" Nov 8 01:39:12.405612 kubelet[3256]: I1108 01:39:12.405595 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cts9h\" (UniqueName: \"kubernetes.io/projected/f5773a4d-1a91-4ae3-82ea-bf0402f8aa01-kube-api-access-cts9h\") pod \"calico-kube-controllers-566b9cf58b-whc7z\" (UID: \"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01\") " pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" Nov 8 01:39:12.406000 kubelet[3256]: I1108 01:39:12.405711 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f0fc8a6-454f-454b-80b2-7ee71f1fc322-config-volume\") pod \"coredns-668d6bf9bc-l4wcp\" (UID: \"4f0fc8a6-454f-454b-80b2-7ee71f1fc322\") " pod="kube-system/coredns-668d6bf9bc-l4wcp" Nov 8 01:39:12.406000 kubelet[3256]: I1108 01:39:12.405773 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f897823-41ac-4b66-8214-3b03dd8d6290-goldmane-ca-bundle\") pod \"goldmane-666569f655-8ld4r\" (UID: \"5f897823-41ac-4b66-8214-3b03dd8d6290\") " pod="calico-system/goldmane-666569f655-8ld4r" Nov 8 01:39:12.406000 kubelet[3256]: I1108 01:39:12.405824 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/555cf867-6266-4755-ac35-32d63c26e55d-whisker-backend-key-pair\") pod \"whisker-5d6f95b8f7-6lzzn\" (UID: \"555cf867-6266-4755-ac35-32d63c26e55d\") " pod="calico-system/whisker-5d6f95b8f7-6lzzn" Nov 8 01:39:12.406000 kubelet[3256]: I1108 01:39:12.405878 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv77z\" (UniqueName: \"kubernetes.io/projected/79b0696d-13d8-44db-9d91-6f1131e9b5ef-kube-api-access-wv77z\") pod \"calico-apiserver-66c4f75c9f-qhkcw\" (UID: \"79b0696d-13d8-44db-9d91-6f1131e9b5ef\") " pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" Nov 8 01:39:12.406000 kubelet[3256]: I1108 01:39:12.405929 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2-config-volume\") pod \"coredns-668d6bf9bc-nsnp6\" (UID: \"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2\") " pod="kube-system/coredns-668d6bf9bc-nsnp6" Nov 8 01:39:12.406841 kubelet[3256]: I1108 01:39:12.405976 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcx8p\" (UniqueName: \"kubernetes.io/projected/555cf867-6266-4755-ac35-32d63c26e55d-kube-api-access-vcx8p\") pod \"whisker-5d6f95b8f7-6lzzn\" (UID: \"555cf867-6266-4755-ac35-32d63c26e55d\") " pod="calico-system/whisker-5d6f95b8f7-6lzzn" Nov 8 01:39:12.406841 kubelet[3256]: I1108 01:39:12.406085 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/79b0696d-13d8-44db-9d91-6f1131e9b5ef-calico-apiserver-certs\") pod \"calico-apiserver-66c4f75c9f-qhkcw\" (UID: \"79b0696d-13d8-44db-9d91-6f1131e9b5ef\") " pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" Nov 8 01:39:12.406841 kubelet[3256]: I1108 01:39:12.406210 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5773a4d-1a91-4ae3-82ea-bf0402f8aa01-tigera-ca-bundle\") pod \"calico-kube-controllers-566b9cf58b-whc7z\" (UID: \"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01\") " pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" Nov 8 01:39:12.406841 kubelet[3256]: I1108 01:39:12.406274 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/31ac439b-7263-4e83-932f-3464165d69ab-calico-apiserver-certs\") pod \"calico-apiserver-66c4f75c9f-gztt5\" (UID: \"31ac439b-7263-4e83-932f-3464165d69ab\") " pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" Nov 8 01:39:12.406841 kubelet[3256]: I1108 01:39:12.406326 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f897823-41ac-4b66-8214-3b03dd8d6290-config\") pod \"goldmane-666569f655-8ld4r\" (UID: \"5f897823-41ac-4b66-8214-3b03dd8d6290\") " pod="calico-system/goldmane-666569f655-8ld4r" Nov 8 01:39:12.407504 kubelet[3256]: I1108 01:39:12.406377 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/555cf867-6266-4755-ac35-32d63c26e55d-whisker-ca-bundle\") pod \"whisker-5d6f95b8f7-6lzzn\" (UID: \"555cf867-6266-4755-ac35-32d63c26e55d\") " pod="calico-system/whisker-5d6f95b8f7-6lzzn" Nov 8 01:39:12.407504 kubelet[3256]: I1108 01:39:12.406431 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b49rq\" (UniqueName: \"kubernetes.io/projected/31ac439b-7263-4e83-932f-3464165d69ab-kube-api-access-b49rq\") pod \"calico-apiserver-66c4f75c9f-gztt5\" (UID: \"31ac439b-7263-4e83-932f-3464165d69ab\") " pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" Nov 8 01:39:12.407504 kubelet[3256]: I1108 01:39:12.406578 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd7k9\" (UniqueName: \"kubernetes.io/projected/4f0fc8a6-454f-454b-80b2-7ee71f1fc322-kube-api-access-qd7k9\") pod \"coredns-668d6bf9bc-l4wcp\" (UID: \"4f0fc8a6-454f-454b-80b2-7ee71f1fc322\") " pod="kube-system/coredns-668d6bf9bc-l4wcp" Nov 8 01:39:12.407504 kubelet[3256]: I1108 01:39:12.406776 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk4w7\" (UniqueName: \"kubernetes.io/projected/2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2-kube-api-access-vk4w7\") pod \"coredns-668d6bf9bc-nsnp6\" (UID: \"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2\") " pod="kube-system/coredns-668d6bf9bc-nsnp6" Nov 8 01:39:12.407504 kubelet[3256]: I1108 01:39:12.406915 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/5f897823-41ac-4b66-8214-3b03dd8d6290-goldmane-key-pair\") pod \"goldmane-666569f655-8ld4r\" (UID: \"5f897823-41ac-4b66-8214-3b03dd8d6290\") " pod="calico-system/goldmane-666569f655-8ld4r" Nov 8 01:39:12.603395 containerd[1926]: time="2025-11-08T01:39:12.603309616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-8ld4r,Uid:5f897823-41ac-4b66-8214-3b03dd8d6290,Namespace:calico-system,Attempt:0,}" Nov 8 01:39:12.604562 containerd[1926]: time="2025-11-08T01:39:12.603305917Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-nsnp6,Uid:2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2,Namespace:kube-system,Attempt:0,}" Nov 8 01:39:12.606593 containerd[1926]: time="2025-11-08T01:39:12.606480823Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l4wcp,Uid:4f0fc8a6-454f-454b-80b2-7ee71f1fc322,Namespace:kube-system,Attempt:0,}" Nov 8 01:39:12.608853 containerd[1926]: time="2025-11-08T01:39:12.608742136Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-gztt5,Uid:31ac439b-7263-4e83-932f-3464165d69ab,Namespace:calico-apiserver,Attempt:0,}" Nov 8 01:39:12.609080 containerd[1926]: time="2025-11-08T01:39:12.608841728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-566b9cf58b-whc7z,Uid:f5773a4d-1a91-4ae3-82ea-bf0402f8aa01,Namespace:calico-system,Attempt:0,}" Nov 8 01:39:12.609522 containerd[1926]: time="2025-11-08T01:39:12.609448198Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-qhkcw,Uid:79b0696d-13d8-44db-9d91-6f1131e9b5ef,Namespace:calico-apiserver,Attempt:0,}" Nov 8 01:39:12.609746 containerd[1926]: time="2025-11-08T01:39:12.609628038Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d6f95b8f7-6lzzn,Uid:555cf867-6266-4755-ac35-32d63c26e55d,Namespace:calico-system,Attempt:0,}" Nov 8 01:39:12.827812 containerd[1926]: time="2025-11-08T01:39:12.827764122Z" level=info msg="shim disconnected" id=d692cb2b654e18a1fa4cd291557e3a065c309a087760dc92a27810c2eef9f723 namespace=k8s.io Nov 8 01:39:12.827908 containerd[1926]: time="2025-11-08T01:39:12.827811532Z" level=warning msg="cleaning up after shim disconnected" id=d692cb2b654e18a1fa4cd291557e3a065c309a087760dc92a27810c2eef9f723 namespace=k8s.io Nov 8 01:39:12.827908 containerd[1926]: time="2025-11-08T01:39:12.827824751Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 8 01:39:12.871970 containerd[1926]: time="2025-11-08T01:39:12.871876283Z" level=error msg="Failed to destroy network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.872252 containerd[1926]: time="2025-11-08T01:39:12.872235664Z" level=error msg="encountered an error cleaning up failed sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.872297 containerd[1926]: time="2025-11-08T01:39:12.872266894Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-8ld4r,Uid:5f897823-41ac-4b66-8214-3b03dd8d6290,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.872447 kubelet[3256]: E1108 01:39:12.872417 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.872636 kubelet[3256]: E1108 01:39:12.872478 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-8ld4r" Nov 8 01:39:12.872636 kubelet[3256]: E1108 01:39:12.872500 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-8ld4r" Nov 8 01:39:12.872636 kubelet[3256]: E1108 01:39:12.872539 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:39:12.873397 containerd[1926]: time="2025-11-08T01:39:12.873373467Z" level=error msg="Failed to destroy network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873565 containerd[1926]: time="2025-11-08T01:39:12.873548798Z" level=error msg="Failed to destroy network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873623 containerd[1926]: time="2025-11-08T01:39:12.873603989Z" level=error msg="encountered an error cleaning up failed sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873658 containerd[1926]: time="2025-11-08T01:39:12.873642472Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-nsnp6,Uid:2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873711 containerd[1926]: time="2025-11-08T01:39:12.873698771Z" level=error msg="encountered an error cleaning up failed sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873731 containerd[1926]: time="2025-11-08T01:39:12.873720566Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-qhkcw,Uid:79b0696d-13d8-44db-9d91-6f1131e9b5ef,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873763 containerd[1926]: time="2025-11-08T01:39:12.873743345Z" level=error msg="Failed to destroy network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873799 kubelet[3256]: E1108 01:39:12.873778 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873843 kubelet[3256]: E1108 01:39:12.873812 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-nsnp6" Nov 8 01:39:12.873843 kubelet[3256]: E1108 01:39:12.873824 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-nsnp6" Nov 8 01:39:12.873884 kubelet[3256]: E1108 01:39:12.873845 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-nsnp6_kube-system(2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-nsnp6_kube-system(2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-nsnp6" podUID="2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2" Nov 8 01:39:12.873884 kubelet[3256]: E1108 01:39:12.873785 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.873884 kubelet[3256]: E1108 01:39:12.873876 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" Nov 8 01:39:12.873958 kubelet[3256]: E1108 01:39:12.873894 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" Nov 8 01:39:12.873958 kubelet[3256]: E1108 01:39:12.873927 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:39:12.874009 containerd[1926]: time="2025-11-08T01:39:12.873883935Z" level=error msg="encountered an error cleaning up failed sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.874009 containerd[1926]: time="2025-11-08T01:39:12.873903539Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l4wcp,Uid:4f0fc8a6-454f-454b-80b2-7ee71f1fc322,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.874063 kubelet[3256]: E1108 01:39:12.873964 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.874046 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34-shm.mount: Deactivated successfully. Nov 8 01:39:12.874192 kubelet[3256]: E1108 01:39:12.874166 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-l4wcp" Nov 8 01:39:12.874192 kubelet[3256]: E1108 01:39:12.874183 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-l4wcp" Nov 8 01:39:12.874260 kubelet[3256]: E1108 01:39:12.874208 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-l4wcp_kube-system(4f0fc8a6-454f-454b-80b2-7ee71f1fc322)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-l4wcp_kube-system(4f0fc8a6-454f-454b-80b2-7ee71f1fc322)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-l4wcp" podUID="4f0fc8a6-454f-454b-80b2-7ee71f1fc322" Nov 8 01:39:12.875129 containerd[1926]: time="2025-11-08T01:39:12.875109293Z" level=error msg="Failed to destroy network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875277 containerd[1926]: time="2025-11-08T01:39:12.875260916Z" level=error msg="Failed to destroy network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875304 containerd[1926]: time="2025-11-08T01:39:12.875280130Z" level=error msg="encountered an error cleaning up failed sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875329 containerd[1926]: time="2025-11-08T01:39:12.875303054Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-566b9cf58b-whc7z,Uid:f5773a4d-1a91-4ae3-82ea-bf0402f8aa01,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875420 kubelet[3256]: E1108 01:39:12.875405 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875444 kubelet[3256]: E1108 01:39:12.875430 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" Nov 8 01:39:12.875464 containerd[1926]: time="2025-11-08T01:39:12.875423081Z" level=error msg="encountered an error cleaning up failed sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875464 containerd[1926]: time="2025-11-08T01:39:12.875446133Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d6f95b8f7-6lzzn,Uid:555cf867-6266-4755-ac35-32d63c26e55d,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875513 kubelet[3256]: E1108 01:39:12.875447 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" Nov 8 01:39:12.875513 kubelet[3256]: E1108 01:39:12.875468 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:39:12.875563 containerd[1926]: time="2025-11-08T01:39:12.875461994Z" level=error msg="Failed to destroy network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875587 kubelet[3256]: E1108 01:39:12.875510 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875587 kubelet[3256]: E1108 01:39:12.875556 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5d6f95b8f7-6lzzn" Nov 8 01:39:12.875587 kubelet[3256]: E1108 01:39:12.875572 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5d6f95b8f7-6lzzn" Nov 8 01:39:12.875648 kubelet[3256]: E1108 01:39:12.875630 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-5d6f95b8f7-6lzzn_calico-system(555cf867-6266-4755-ac35-32d63c26e55d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-5d6f95b8f7-6lzzn_calico-system(555cf867-6266-4755-ac35-32d63c26e55d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5d6f95b8f7-6lzzn" podUID="555cf867-6266-4755-ac35-32d63c26e55d" Nov 8 01:39:12.875681 containerd[1926]: time="2025-11-08T01:39:12.875601377Z" level=error msg="encountered an error cleaning up failed sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875681 containerd[1926]: time="2025-11-08T01:39:12.875621905Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-gztt5,Uid:31ac439b-7263-4e83-932f-3464165d69ab,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875727 kubelet[3256]: E1108 01:39:12.875681 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:12.875727 kubelet[3256]: E1108 01:39:12.875697 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" Nov 8 01:39:12.875727 kubelet[3256]: E1108 01:39:12.875707 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" Nov 8 01:39:12.875783 kubelet[3256]: E1108 01:39:12.875722 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:39:13.589776 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da-shm.mount: Deactivated successfully. Nov 8 01:39:13.589856 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d-shm.mount: Deactivated successfully. Nov 8 01:39:13.589909 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c-shm.mount: Deactivated successfully. Nov 8 01:39:13.589964 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6-shm.mount: Deactivated successfully. Nov 8 01:39:13.590016 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900-shm.mount: Deactivated successfully. Nov 8 01:39:13.590066 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff-shm.mount: Deactivated successfully. Nov 8 01:39:13.716769 containerd[1926]: time="2025-11-08T01:39:13.716708111Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dmqbg,Uid:ff4b01f8-4bf6-4050-afe4-e2ac586eece3,Namespace:calico-system,Attempt:0,}" Nov 8 01:39:13.743997 containerd[1926]: time="2025-11-08T01:39:13.743949854Z" level=error msg="Failed to destroy network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.744224 containerd[1926]: time="2025-11-08T01:39:13.744145119Z" level=error msg="encountered an error cleaning up failed sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.744224 containerd[1926]: time="2025-11-08T01:39:13.744222109Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dmqbg,Uid:ff4b01f8-4bf6-4050-afe4-e2ac586eece3,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.744443 kubelet[3256]: E1108 01:39:13.744390 3256 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.744443 kubelet[3256]: E1108 01:39:13.744430 3256 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-dmqbg" Nov 8 01:39:13.744505 kubelet[3256]: E1108 01:39:13.744444 3256 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-dmqbg" Nov 8 01:39:13.744505 kubelet[3256]: E1108 01:39:13.744472 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:13.745724 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11-shm.mount: Deactivated successfully. Nov 8 01:39:13.790744 kubelet[3256]: I1108 01:39:13.790675 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:13.791370 containerd[1926]: time="2025-11-08T01:39:13.791294971Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Nov 8 01:39:13.792221 containerd[1926]: time="2025-11-08T01:39:13.792207852Z" level=info msg="StopPodSandbox for \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\"" Nov 8 01:39:13.792320 containerd[1926]: time="2025-11-08T01:39:13.792308736Z" level=info msg="Ensure that sandbox 0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d in task-service has been cleanup successfully" Nov 8 01:39:13.792368 kubelet[3256]: I1108 01:39:13.792358 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:13.792629 containerd[1926]: time="2025-11-08T01:39:13.792615544Z" level=info msg="StopPodSandbox for \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\"" Nov 8 01:39:13.792724 containerd[1926]: time="2025-11-08T01:39:13.792712456Z" level=info msg="Ensure that sandbox 4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900 in task-service has been cleanup successfully" Nov 8 01:39:13.792799 kubelet[3256]: I1108 01:39:13.792789 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:13.793024 containerd[1926]: time="2025-11-08T01:39:13.793011379Z" level=info msg="StopPodSandbox for \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\"" Nov 8 01:39:13.793127 containerd[1926]: time="2025-11-08T01:39:13.793116565Z" level=info msg="Ensure that sandbox 45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff in task-service has been cleanup successfully" Nov 8 01:39:13.793212 kubelet[3256]: I1108 01:39:13.793204 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:13.793451 containerd[1926]: time="2025-11-08T01:39:13.793441246Z" level=info msg="StopPodSandbox for \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\"" Nov 8 01:39:13.793522 containerd[1926]: time="2025-11-08T01:39:13.793512873Z" level=info msg="Ensure that sandbox 7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da in task-service has been cleanup successfully" Nov 8 01:39:13.793727 kubelet[3256]: I1108 01:39:13.793712 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:13.794022 containerd[1926]: time="2025-11-08T01:39:13.794006867Z" level=info msg="StopPodSandbox for \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\"" Nov 8 01:39:13.794188 containerd[1926]: time="2025-11-08T01:39:13.794172318Z" level=info msg="Ensure that sandbox e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c in task-service has been cleanup successfully" Nov 8 01:39:13.794331 kubelet[3256]: I1108 01:39:13.794320 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:13.794688 containerd[1926]: time="2025-11-08T01:39:13.794670944Z" level=info msg="StopPodSandbox for \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\"" Nov 8 01:39:13.794799 containerd[1926]: time="2025-11-08T01:39:13.794789152Z" level=info msg="Ensure that sandbox 1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6 in task-service has been cleanup successfully" Nov 8 01:39:13.794971 kubelet[3256]: I1108 01:39:13.794960 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:13.795416 containerd[1926]: time="2025-11-08T01:39:13.795389283Z" level=info msg="StopPodSandbox for \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\"" Nov 8 01:39:13.795978 kubelet[3256]: I1108 01:39:13.795958 3256 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:13.796109 containerd[1926]: time="2025-11-08T01:39:13.795931874Z" level=info msg="Ensure that sandbox a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11 in task-service has been cleanup successfully" Nov 8 01:39:13.797303 containerd[1926]: time="2025-11-08T01:39:13.796664165Z" level=info msg="StopPodSandbox for \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\"" Nov 8 01:39:13.797303 containerd[1926]: time="2025-11-08T01:39:13.796872831Z" level=info msg="Ensure that sandbox 38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34 in task-service has been cleanup successfully" Nov 8 01:39:13.811902 containerd[1926]: time="2025-11-08T01:39:13.811859987Z" level=error msg="StopPodSandbox for \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\" failed" error="failed to destroy network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.812085 kubelet[3256]: E1108 01:39:13.812053 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:13.812152 kubelet[3256]: E1108 01:39:13.812113 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff"} Nov 8 01:39:13.812188 kubelet[3256]: E1108 01:39:13.812174 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.812262 kubelet[3256]: E1108 01:39:13.812194 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:39:13.812321 containerd[1926]: time="2025-11-08T01:39:13.812233689Z" level=error msg="StopPodSandbox for \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\" failed" error="failed to destroy network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.812395 kubelet[3256]: E1108 01:39:13.812301 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:13.812395 kubelet[3256]: E1108 01:39:13.812318 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d"} Nov 8 01:39:13.812395 kubelet[3256]: E1108 01:39:13.812334 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"555cf867-6266-4755-ac35-32d63c26e55d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.812395 kubelet[3256]: E1108 01:39:13.812345 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"555cf867-6266-4755-ac35-32d63c26e55d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5d6f95b8f7-6lzzn" podUID="555cf867-6266-4755-ac35-32d63c26e55d" Nov 8 01:39:13.812695 containerd[1926]: time="2025-11-08T01:39:13.812676334Z" level=error msg="StopPodSandbox for \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\" failed" error="failed to destroy network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.812761 kubelet[3256]: E1108 01:39:13.812745 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:13.812801 kubelet[3256]: E1108 01:39:13.812767 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900"} Nov 8 01:39:13.812801 kubelet[3256]: E1108 01:39:13.812786 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.812884 kubelet[3256]: E1108 01:39:13.812804 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-nsnp6" podUID="2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2" Nov 8 01:39:13.813084 containerd[1926]: time="2025-11-08T01:39:13.813061184Z" level=error msg="StopPodSandbox for \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\" failed" error="failed to destroy network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.813159 kubelet[3256]: E1108 01:39:13.813134 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:13.813202 kubelet[3256]: E1108 01:39:13.813163 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6"} Nov 8 01:39:13.813202 kubelet[3256]: E1108 01:39:13.813179 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"4f0fc8a6-454f-454b-80b2-7ee71f1fc322\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.813202 kubelet[3256]: E1108 01:39:13.813195 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"4f0fc8a6-454f-454b-80b2-7ee71f1fc322\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-l4wcp" podUID="4f0fc8a6-454f-454b-80b2-7ee71f1fc322" Nov 8 01:39:13.813333 containerd[1926]: time="2025-11-08T01:39:13.813320455Z" level=error msg="StopPodSandbox for \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\" failed" error="failed to destroy network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.813391 kubelet[3256]: E1108 01:39:13.813382 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:13.813411 kubelet[3256]: E1108 01:39:13.813396 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c"} Nov 8 01:39:13.813428 kubelet[3256]: E1108 01:39:13.813409 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"31ac439b-7263-4e83-932f-3464165d69ab\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.813428 kubelet[3256]: E1108 01:39:13.813420 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"31ac439b-7263-4e83-932f-3464165d69ab\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:39:13.813487 containerd[1926]: time="2025-11-08T01:39:13.813454410Z" level=error msg="StopPodSandbox for \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\" failed" error="failed to destroy network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.813595 kubelet[3256]: E1108 01:39:13.813538 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:13.813595 kubelet[3256]: E1108 01:39:13.813555 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34"} Nov 8 01:39:13.813595 kubelet[3256]: E1108 01:39:13.813569 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"5f897823-41ac-4b66-8214-3b03dd8d6290\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.813595 kubelet[3256]: E1108 01:39:13.813581 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"5f897823-41ac-4b66-8214-3b03dd8d6290\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:39:13.813712 containerd[1926]: time="2025-11-08T01:39:13.813552813Z" level=error msg="StopPodSandbox for \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\" failed" error="failed to destroy network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.813735 kubelet[3256]: E1108 01:39:13.813606 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:13.813735 kubelet[3256]: E1108 01:39:13.813617 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11"} Nov 8 01:39:13.813735 kubelet[3256]: E1108 01:39:13.813629 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ff4b01f8-4bf6-4050-afe4-e2ac586eece3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.813735 kubelet[3256]: E1108 01:39:13.813639 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ff4b01f8-4bf6-4050-afe4-e2ac586eece3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:13.813824 containerd[1926]: time="2025-11-08T01:39:13.813772767Z" level=error msg="StopPodSandbox for \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\" failed" error="failed to destroy network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Nov 8 01:39:13.813844 kubelet[3256]: E1108 01:39:13.813833 3256 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:13.813865 kubelet[3256]: E1108 01:39:13.813844 3256 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da"} Nov 8 01:39:13.813865 kubelet[3256]: E1108 01:39:13.813855 3256 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"79b0696d-13d8-44db-9d91-6f1131e9b5ef\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Nov 8 01:39:13.813913 kubelet[3256]: E1108 01:39:13.813863 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"79b0696d-13d8-44db-9d91-6f1131e9b5ef\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:39:16.986769 kubelet[3256]: I1108 01:39:16.986704 3256 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 8 01:39:18.558178 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount329008048.mount: Deactivated successfully. Nov 8 01:39:18.574668 containerd[1926]: time="2025-11-08T01:39:18.574620166Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:18.574833 containerd[1926]: time="2025-11-08T01:39:18.574819482Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156883675" Nov 8 01:39:18.575217 containerd[1926]: time="2025-11-08T01:39:18.575179613Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:18.576068 containerd[1926]: time="2025-11-08T01:39:18.576053468Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 8 01:39:18.576467 containerd[1926]: time="2025-11-08T01:39:18.576452083Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 4.78508226s" Nov 8 01:39:18.576511 containerd[1926]: time="2025-11-08T01:39:18.576470676Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Nov 8 01:39:18.579761 containerd[1926]: time="2025-11-08T01:39:18.579744067Z" level=info msg="CreateContainer within sandbox \"0e8517d4481369d52be9064628ad07d03821e659106ff00999716de429622f6d\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Nov 8 01:39:18.585836 containerd[1926]: time="2025-11-08T01:39:18.585787251Z" level=info msg="CreateContainer within sandbox \"0e8517d4481369d52be9064628ad07d03821e659106ff00999716de429622f6d\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"e8ad0ba3e6560f7fcffcd6d3eeffc3af49a9157b8b7f8235c174e662c431b02d\"" Nov 8 01:39:18.586046 containerd[1926]: time="2025-11-08T01:39:18.586037094Z" level=info msg="StartContainer for \"e8ad0ba3e6560f7fcffcd6d3eeffc3af49a9157b8b7f8235c174e662c431b02d\"" Nov 8 01:39:18.629427 containerd[1926]: time="2025-11-08T01:39:18.629376423Z" level=info msg="StartContainer for \"e8ad0ba3e6560f7fcffcd6d3eeffc3af49a9157b8b7f8235c174e662c431b02d\" returns successfully" Nov 8 01:39:18.708194 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Nov 8 01:39:18.708253 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Nov 8 01:39:18.749615 containerd[1926]: time="2025-11-08T01:39:18.749581896Z" level=info msg="StopPodSandbox for \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\"" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.775 [INFO][4760] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.775 [INFO][4760] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" iface="eth0" netns="/var/run/netns/cni-a08af884-7bd3-1c9a-9e88-d69d9d97c631" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.775 [INFO][4760] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" iface="eth0" netns="/var/run/netns/cni-a08af884-7bd3-1c9a-9e88-d69d9d97c631" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.775 [INFO][4760] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" iface="eth0" netns="/var/run/netns/cni-a08af884-7bd3-1c9a-9e88-d69d9d97c631" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.775 [INFO][4760] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.775 [INFO][4760] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.787 [INFO][4788] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.788 [INFO][4788] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.788 [INFO][4788] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.791 [WARNING][4788] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.791 [INFO][4788] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.792 [INFO][4788] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:18.795036 containerd[1926]: 2025-11-08 01:39:18.793 [INFO][4760] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:18.795421 containerd[1926]: time="2025-11-08T01:39:18.795123988Z" level=info msg="TearDown network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\" successfully" Nov 8 01:39:18.795421 containerd[1926]: time="2025-11-08T01:39:18.795145036Z" level=info msg="StopPodSandbox for \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\" returns successfully" Nov 8 01:39:18.828912 kubelet[3256]: I1108 01:39:18.828824 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-6fmld" podStartSLOduration=1.2338755670000001 podStartE2EDuration="15.828811529s" podCreationTimestamp="2025-11-08 01:39:03 +0000 UTC" firstStartedPulling="2025-11-08 01:39:03.981891772 +0000 UTC m=+18.314072518" lastFinishedPulling="2025-11-08 01:39:18.576827736 +0000 UTC m=+32.909008480" observedRunningTime="2025-11-08 01:39:18.828630494 +0000 UTC m=+33.160811241" watchObservedRunningTime="2025-11-08 01:39:18.828811529 +0000 UTC m=+33.160992273" Nov 8 01:39:18.848490 kubelet[3256]: I1108 01:39:18.848462 3256 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcx8p\" (UniqueName: \"kubernetes.io/projected/555cf867-6266-4755-ac35-32d63c26e55d-kube-api-access-vcx8p\") pod \"555cf867-6266-4755-ac35-32d63c26e55d\" (UID: \"555cf867-6266-4755-ac35-32d63c26e55d\") " Nov 8 01:39:18.848592 kubelet[3256]: I1108 01:39:18.848511 3256 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/555cf867-6266-4755-ac35-32d63c26e55d-whisker-backend-key-pair\") pod \"555cf867-6266-4755-ac35-32d63c26e55d\" (UID: \"555cf867-6266-4755-ac35-32d63c26e55d\") " Nov 8 01:39:18.848592 kubelet[3256]: I1108 01:39:18.848537 3256 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/555cf867-6266-4755-ac35-32d63c26e55d-whisker-ca-bundle\") pod \"555cf867-6266-4755-ac35-32d63c26e55d\" (UID: \"555cf867-6266-4755-ac35-32d63c26e55d\") " Nov 8 01:39:18.848998 kubelet[3256]: I1108 01:39:18.848976 3256 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/555cf867-6266-4755-ac35-32d63c26e55d-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "555cf867-6266-4755-ac35-32d63c26e55d" (UID: "555cf867-6266-4755-ac35-32d63c26e55d"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Nov 8 01:39:18.850113 kubelet[3256]: I1108 01:39:18.850092 3256 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555cf867-6266-4755-ac35-32d63c26e55d-kube-api-access-vcx8p" (OuterVolumeSpecName: "kube-api-access-vcx8p") pod "555cf867-6266-4755-ac35-32d63c26e55d" (UID: "555cf867-6266-4755-ac35-32d63c26e55d"). InnerVolumeSpecName "kube-api-access-vcx8p". PluginName "kubernetes.io/projected", VolumeGIDValue "" Nov 8 01:39:18.850113 kubelet[3256]: I1108 01:39:18.850099 3256 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/555cf867-6266-4755-ac35-32d63c26e55d-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "555cf867-6266-4755-ac35-32d63c26e55d" (UID: "555cf867-6266-4755-ac35-32d63c26e55d"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Nov 8 01:39:18.949719 kubelet[3256]: I1108 01:39:18.949610 3256 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vcx8p\" (UniqueName: \"kubernetes.io/projected/555cf867-6266-4755-ac35-32d63c26e55d-kube-api-access-vcx8p\") on node \"ci-4081.3.6-n-3863d1ecf5\" DevicePath \"\"" Nov 8 01:39:18.949719 kubelet[3256]: I1108 01:39:18.949682 3256 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/555cf867-6266-4755-ac35-32d63c26e55d-whisker-backend-key-pair\") on node \"ci-4081.3.6-n-3863d1ecf5\" DevicePath \"\"" Nov 8 01:39:18.949719 kubelet[3256]: I1108 01:39:18.949712 3256 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/555cf867-6266-4755-ac35-32d63c26e55d-whisker-ca-bundle\") on node \"ci-4081.3.6-n-3863d1ecf5\" DevicePath \"\"" Nov 8 01:39:19.252274 kubelet[3256]: I1108 01:39:19.252057 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqv8n\" (UniqueName: \"kubernetes.io/projected/8ffe5d99-6a33-4ea1-b47d-a065e9dc4970-kube-api-access-gqv8n\") pod \"whisker-5c9bb6b6fc-xlbx8\" (UID: \"8ffe5d99-6a33-4ea1-b47d-a065e9dc4970\") " pod="calico-system/whisker-5c9bb6b6fc-xlbx8" Nov 8 01:39:19.252274 kubelet[3256]: I1108 01:39:19.252246 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/8ffe5d99-6a33-4ea1-b47d-a065e9dc4970-whisker-backend-key-pair\") pod \"whisker-5c9bb6b6fc-xlbx8\" (UID: \"8ffe5d99-6a33-4ea1-b47d-a065e9dc4970\") " pod="calico-system/whisker-5c9bb6b6fc-xlbx8" Nov 8 01:39:19.252607 kubelet[3256]: I1108 01:39:19.252333 3256 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ffe5d99-6a33-4ea1-b47d-a065e9dc4970-whisker-ca-bundle\") pod \"whisker-5c9bb6b6fc-xlbx8\" (UID: \"8ffe5d99-6a33-4ea1-b47d-a065e9dc4970\") " pod="calico-system/whisker-5c9bb6b6fc-xlbx8" Nov 8 01:39:19.475734 containerd[1926]: time="2025-11-08T01:39:19.475625826Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5c9bb6b6fc-xlbx8,Uid:8ffe5d99-6a33-4ea1-b47d-a065e9dc4970,Namespace:calico-system,Attempt:0,}" Nov 8 01:39:19.537623 systemd-networkd[1553]: cali936859659ef: Link UP Nov 8 01:39:19.537732 systemd-networkd[1553]: cali936859659ef: Gained carrier Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.491 [INFO][4844] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.498 [INFO][4844] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0 whisker-5c9bb6b6fc- calico-system 8ffe5d99-6a33-4ea1-b47d-a065e9dc4970 899 0 2025-11-08 01:39:19 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:5c9bb6b6fc projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 whisker-5c9bb6b6fc-xlbx8 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali936859659ef [] [] }} ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.498 [INFO][4844] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.511 [INFO][4865] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" HandleID="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.512 [INFO][4865] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" HandleID="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002cf5f0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"whisker-5c9bb6b6fc-xlbx8", "timestamp":"2025-11-08 01:39:19.511962586 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.512 [INFO][4865] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.512 [INFO][4865] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.512 [INFO][4865] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.516 [INFO][4865] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.519 [INFO][4865] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.522 [INFO][4865] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.524 [INFO][4865] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.525 [INFO][4865] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.525 [INFO][4865] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.527 [INFO][4865] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.529 [INFO][4865] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.532 [INFO][4865] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.65/26] block=192.168.68.64/26 handle="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.532 [INFO][4865] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.65/26] handle="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.532 [INFO][4865] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:19.543764 containerd[1926]: 2025-11-08 01:39:19.532 [INFO][4865] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.65/26] IPv6=[] ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" HandleID="k8s-pod-network.ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" Nov 8 01:39:19.544219 containerd[1926]: 2025-11-08 01:39:19.533 [INFO][4844] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0", GenerateName:"whisker-5c9bb6b6fc-", Namespace:"calico-system", SelfLink:"", UID:"8ffe5d99-6a33-4ea1-b47d-a065e9dc4970", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5c9bb6b6fc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"whisker-5c9bb6b6fc-xlbx8", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.68.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali936859659ef", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:19.544219 containerd[1926]: 2025-11-08 01:39:19.533 [INFO][4844] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.65/32] ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" Nov 8 01:39:19.544219 containerd[1926]: 2025-11-08 01:39:19.533 [INFO][4844] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali936859659ef ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" Nov 8 01:39:19.544219 containerd[1926]: 2025-11-08 01:39:19.537 [INFO][4844] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" Nov 8 01:39:19.544219 containerd[1926]: 2025-11-08 01:39:19.538 [INFO][4844] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0", GenerateName:"whisker-5c9bb6b6fc-", Namespace:"calico-system", SelfLink:"", UID:"8ffe5d99-6a33-4ea1-b47d-a065e9dc4970", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5c9bb6b6fc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c", Pod:"whisker-5c9bb6b6fc-xlbx8", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.68.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali936859659ef", MAC:"aa:95:59:c8:7f:c0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:19.544219 containerd[1926]: 2025-11-08 01:39:19.542 [INFO][4844] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c" Namespace="calico-system" Pod="whisker-5c9bb6b6fc-xlbx8" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5c9bb6b6fc--xlbx8-eth0" Nov 8 01:39:19.552228 containerd[1926]: time="2025-11-08T01:39:19.552130069Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:19.552228 containerd[1926]: time="2025-11-08T01:39:19.552164935Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:19.552228 containerd[1926]: time="2025-11-08T01:39:19.552172145Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:19.552228 containerd[1926]: time="2025-11-08T01:39:19.552213956Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:19.561117 systemd[1]: run-netns-cni\x2da08af884\x2d7bd3\x2d1c9a\x2d9e88\x2dd69d9d97c631.mount: Deactivated successfully. Nov 8 01:39:19.561197 systemd[1]: var-lib-kubelet-pods-555cf867\x2d6266\x2d4755\x2dac35\x2d32d63c26e55d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dvcx8p.mount: Deactivated successfully. Nov 8 01:39:19.561255 systemd[1]: var-lib-kubelet-pods-555cf867\x2d6266\x2d4755\x2dac35\x2d32d63c26e55d-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Nov 8 01:39:19.591458 containerd[1926]: time="2025-11-08T01:39:19.591414781Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5c9bb6b6fc-xlbx8,Uid:8ffe5d99-6a33-4ea1-b47d-a065e9dc4970,Namespace:calico-system,Attempt:0,} returns sandbox id \"ee65b362da76f37c5d214826d8b7f19dc64fb828e4d1c44913bf94a3fa0b011c\"" Nov 8 01:39:19.592505 containerd[1926]: time="2025-11-08T01:39:19.592278597Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 8 01:39:19.716544 kubelet[3256]: I1108 01:39:19.716448 3256 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="555cf867-6266-4755-ac35-32d63c26e55d" path="/var/lib/kubelet/pods/555cf867-6266-4755-ac35-32d63c26e55d/volumes" Nov 8 01:39:19.938193 kernel: bpftool[5121]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Nov 8 01:39:20.013480 containerd[1926]: time="2025-11-08T01:39:20.013457957Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:20.014024 containerd[1926]: time="2025-11-08T01:39:20.013999707Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 8 01:39:20.014063 containerd[1926]: time="2025-11-08T01:39:20.014026345Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 8 01:39:20.014157 kubelet[3256]: E1108 01:39:20.014121 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:39:20.014338 kubelet[3256]: E1108 01:39:20.014169 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:39:20.014359 kubelet[3256]: E1108 01:39:20.014251 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:52eaa7a533b1456aa7b1871d4874688d,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:20.015766 containerd[1926]: time="2025-11-08T01:39:20.015750024Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 8 01:39:20.097016 systemd-networkd[1553]: vxlan.calico: Link UP Nov 8 01:39:20.097019 systemd-networkd[1553]: vxlan.calico: Gained carrier Nov 8 01:39:20.381762 containerd[1926]: time="2025-11-08T01:39:20.381678435Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:20.382574 containerd[1926]: time="2025-11-08T01:39:20.382491964Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 8 01:39:20.382574 containerd[1926]: time="2025-11-08T01:39:20.382548186Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 8 01:39:20.382713 kubelet[3256]: E1108 01:39:20.382657 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:39:20.382713 kubelet[3256]: E1108 01:39:20.382691 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:39:20.382815 kubelet[3256]: E1108 01:39:20.382756 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:20.384391 kubelet[3256]: E1108 01:39:20.384349 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:39:20.817974 kubelet[3256]: E1108 01:39:20.817812 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:39:21.185404 systemd-networkd[1553]: vxlan.calico: Gained IPv6LL Nov 8 01:39:21.505434 systemd-networkd[1553]: cali936859659ef: Gained IPv6LL Nov 8 01:39:24.711573 containerd[1926]: time="2025-11-08T01:39:24.711491733Z" level=info msg="StopPodSandbox for \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\"" Nov 8 01:39:24.711573 containerd[1926]: time="2025-11-08T01:39:24.711515441Z" level=info msg="StopPodSandbox for \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\"" Nov 8 01:39:24.711573 containerd[1926]: time="2025-11-08T01:39:24.711492428Z" level=info msg="StopPodSandbox for \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\"" Nov 8 01:39:24.711875 containerd[1926]: time="2025-11-08T01:39:24.711491816Z" level=info msg="StopPodSandbox for \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\"" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.733 [INFO][5284] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.734 [INFO][5284] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" iface="eth0" netns="/var/run/netns/cni-7b4ebf39-1573-5804-e2e9-680cb8c62020" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.734 [INFO][5284] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" iface="eth0" netns="/var/run/netns/cni-7b4ebf39-1573-5804-e2e9-680cb8c62020" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.734 [INFO][5284] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" iface="eth0" netns="/var/run/netns/cni-7b4ebf39-1573-5804-e2e9-680cb8c62020" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.734 [INFO][5284] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.734 [INFO][5284] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.753 [INFO][5342] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.753 [INFO][5342] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.753 [INFO][5342] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.756 [WARNING][5342] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.756 [INFO][5342] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.757 [INFO][5342] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:24.758872 containerd[1926]: 2025-11-08 01:39:24.758 [INFO][5284] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:24.759156 containerd[1926]: time="2025-11-08T01:39:24.758921494Z" level=info msg="TearDown network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\" successfully" Nov 8 01:39:24.759156 containerd[1926]: time="2025-11-08T01:39:24.758941269Z" level=info msg="StopPodSandbox for \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\" returns successfully" Nov 8 01:39:24.759382 containerd[1926]: time="2025-11-08T01:39:24.759367128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-qhkcw,Uid:79b0696d-13d8-44db-9d91-6f1131e9b5ef,Namespace:calico-apiserver,Attempt:1,}" Nov 8 01:39:24.761121 systemd[1]: run-netns-cni\x2d7b4ebf39\x2d1573\x2d5804\x2de2e9\x2d680cb8c62020.mount: Deactivated successfully. Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.739 [INFO][5282] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.739 [INFO][5282] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" iface="eth0" netns="/var/run/netns/cni-36c003ce-e715-c728-fb07-2a13c9d6bd8b" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.743 [INFO][5282] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" iface="eth0" netns="/var/run/netns/cni-36c003ce-e715-c728-fb07-2a13c9d6bd8b" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.743 [INFO][5282] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" iface="eth0" netns="/var/run/netns/cni-36c003ce-e715-c728-fb07-2a13c9d6bd8b" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.743 [INFO][5282] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.743 [INFO][5282] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.754 [INFO][5355] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.754 [INFO][5355] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.757 [INFO][5355] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.763 [WARNING][5355] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.763 [INFO][5355] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.764 [INFO][5355] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:24.765831 containerd[1926]: 2025-11-08 01:39:24.764 [INFO][5282] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:24.766158 containerd[1926]: time="2025-11-08T01:39:24.765916715Z" level=info msg="TearDown network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\" successfully" Nov 8 01:39:24.766158 containerd[1926]: time="2025-11-08T01:39:24.765937435Z" level=info msg="StopPodSandbox for \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\" returns successfully" Nov 8 01:39:24.766354 containerd[1926]: time="2025-11-08T01:39:24.766342260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-nsnp6,Uid:2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2,Namespace:kube-system,Attempt:1,}" Nov 8 01:39:24.769983 systemd[1]: run-netns-cni\x2d36c003ce\x2de715\x2dc728\x2dfb07\x2d2a13c9d6bd8b.mount: Deactivated successfully. Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.739 [INFO][5281] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.739 [INFO][5281] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" iface="eth0" netns="/var/run/netns/cni-c4330a83-1126-ff65-24fd-7540bb8a1abc" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.744 [INFO][5281] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" iface="eth0" netns="/var/run/netns/cni-c4330a83-1126-ff65-24fd-7540bb8a1abc" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.744 [INFO][5281] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" iface="eth0" netns="/var/run/netns/cni-c4330a83-1126-ff65-24fd-7540bb8a1abc" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.744 [INFO][5281] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.744 [INFO][5281] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.754 [INFO][5361] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.754 [INFO][5361] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.764 [INFO][5361] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.767 [WARNING][5361] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.767 [INFO][5361] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.768 [INFO][5361] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:24.770071 containerd[1926]: 2025-11-08 01:39:24.769 [INFO][5281] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:24.770304 containerd[1926]: time="2025-11-08T01:39:24.770130184Z" level=info msg="TearDown network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\" successfully" Nov 8 01:39:24.770304 containerd[1926]: time="2025-11-08T01:39:24.770194370Z" level=info msg="StopPodSandbox for \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\" returns successfully" Nov 8 01:39:24.770706 containerd[1926]: time="2025-11-08T01:39:24.770690136Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-566b9cf58b-whc7z,Uid:f5773a4d-1a91-4ae3-82ea-bf0402f8aa01,Namespace:calico-system,Attempt:1,}" Nov 8 01:39:24.772279 systemd[1]: run-netns-cni\x2dc4330a83\x2d1126\x2dff65\x2d24fd\x2d7540bb8a1abc.mount: Deactivated successfully. Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.735 [INFO][5283] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.735 [INFO][5283] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" iface="eth0" netns="/var/run/netns/cni-2ac0e4cf-3444-456d-2c80-92512319f39d" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.735 [INFO][5283] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" iface="eth0" netns="/var/run/netns/cni-2ac0e4cf-3444-456d-2c80-92512319f39d" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.735 [INFO][5283] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" iface="eth0" netns="/var/run/netns/cni-2ac0e4cf-3444-456d-2c80-92512319f39d" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.735 [INFO][5283] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.735 [INFO][5283] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.756 [INFO][5344] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.756 [INFO][5344] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.768 [INFO][5344] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.771 [WARNING][5344] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.772 [INFO][5344] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.772 [INFO][5344] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:24.774264 containerd[1926]: 2025-11-08 01:39:24.773 [INFO][5283] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:24.774512 containerd[1926]: time="2025-11-08T01:39:24.774339321Z" level=info msg="TearDown network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\" successfully" Nov 8 01:39:24.774512 containerd[1926]: time="2025-11-08T01:39:24.774353210Z" level=info msg="StopPodSandbox for \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\" returns successfully" Nov 8 01:39:24.774734 containerd[1926]: time="2025-11-08T01:39:24.774723581Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l4wcp,Uid:4f0fc8a6-454f-454b-80b2-7ee71f1fc322,Namespace:kube-system,Attempt:1,}" Nov 8 01:39:24.822135 systemd-networkd[1553]: calife27e0b4819: Link UP Nov 8 01:39:24.822774 systemd-networkd[1553]: calife27e0b4819: Gained carrier Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.781 [INFO][5392] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0 calico-apiserver-66c4f75c9f- calico-apiserver 79b0696d-13d8-44db-9d91-6f1131e9b5ef 934 0 2025-11-08 01:38:58 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:66c4f75c9f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 calico-apiserver-66c4f75c9f-qhkcw eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calife27e0b4819 [] [] }} ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.781 [INFO][5392] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.795 [INFO][5416] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" HandleID="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.795 [INFO][5416] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" HandleID="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00061b700), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"calico-apiserver-66c4f75c9f-qhkcw", "timestamp":"2025-11-08 01:39:24.795080378 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.795 [INFO][5416] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.795 [INFO][5416] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.795 [INFO][5416] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.801 [INFO][5416] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.804 [INFO][5416] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.809 [INFO][5416] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.810 [INFO][5416] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.812 [INFO][5416] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.812 [INFO][5416] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.813 [INFO][5416] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5 Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.815 [INFO][5416] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.819 [INFO][5416] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.66/26] block=192.168.68.64/26 handle="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.819 [INFO][5416] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.66/26] handle="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.819 [INFO][5416] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:24.830048 containerd[1926]: 2025-11-08 01:39:24.819 [INFO][5416] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.66/26] IPv6=[] ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" HandleID="k8s-pod-network.1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.830453 containerd[1926]: 2025-11-08 01:39:24.820 [INFO][5392] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"79b0696d-13d8-44db-9d91-6f1131e9b5ef", ResourceVersion:"934", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"calico-apiserver-66c4f75c9f-qhkcw", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calife27e0b4819", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:24.830453 containerd[1926]: 2025-11-08 01:39:24.820 [INFO][5392] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.66/32] ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.830453 containerd[1926]: 2025-11-08 01:39:24.821 [INFO][5392] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calife27e0b4819 ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.830453 containerd[1926]: 2025-11-08 01:39:24.822 [INFO][5392] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.830453 containerd[1926]: 2025-11-08 01:39:24.823 [INFO][5392] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"79b0696d-13d8-44db-9d91-6f1131e9b5ef", ResourceVersion:"934", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5", Pod:"calico-apiserver-66c4f75c9f-qhkcw", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calife27e0b4819", MAC:"06:fd:2b:eb:60:c5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:24.830453 containerd[1926]: 2025-11-08 01:39:24.829 [INFO][5392] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-qhkcw" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:24.838087 containerd[1926]: time="2025-11-08T01:39:24.838011982Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:24.838087 containerd[1926]: time="2025-11-08T01:39:24.838046775Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:24.838308 containerd[1926]: time="2025-11-08T01:39:24.838247014Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:24.838308 containerd[1926]: time="2025-11-08T01:39:24.838300665Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:24.897320 containerd[1926]: time="2025-11-08T01:39:24.897298946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-qhkcw,Uid:79b0696d-13d8-44db-9d91-6f1131e9b5ef,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5\"" Nov 8 01:39:24.897993 containerd[1926]: time="2025-11-08T01:39:24.897980813Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:39:24.916925 systemd-networkd[1553]: cali3403d4852c5: Link UP Nov 8 01:39:24.917071 systemd-networkd[1553]: cali3403d4852c5: Gained carrier Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.804 [INFO][5421] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0 coredns-668d6bf9bc- kube-system 2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2 935 0 2025-11-08 01:38:51 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 coredns-668d6bf9bc-nsnp6 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali3403d4852c5 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.804 [INFO][5421] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.821 [INFO][5490] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" HandleID="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.821 [INFO][5490] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" HandleID="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000345ae0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"coredns-668d6bf9bc-nsnp6", "timestamp":"2025-11-08 01:39:24.821652065 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.821 [INFO][5490] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.821 [INFO][5490] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.821 [INFO][5490] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.902 [INFO][5490] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.905 [INFO][5490] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.908 [INFO][5490] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.908 [INFO][5490] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.909 [INFO][5490] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.909 [INFO][5490] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.910 [INFO][5490] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4 Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.912 [INFO][5490] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.915 [INFO][5490] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.67/26] block=192.168.68.64/26 handle="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.915 [INFO][5490] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.67/26] handle="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.915 [INFO][5490] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:24.922221 containerd[1926]: 2025-11-08 01:39:24.915 [INFO][5490] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.67/26] IPv6=[] ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" HandleID="k8s-pod-network.6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.922607 containerd[1926]: 2025-11-08 01:39:24.916 [INFO][5421] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2", ResourceVersion:"935", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"coredns-668d6bf9bc-nsnp6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3403d4852c5", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:24.922607 containerd[1926]: 2025-11-08 01:39:24.916 [INFO][5421] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.67/32] ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.922607 containerd[1926]: 2025-11-08 01:39:24.916 [INFO][5421] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3403d4852c5 ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.922607 containerd[1926]: 2025-11-08 01:39:24.917 [INFO][5421] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.922607 containerd[1926]: 2025-11-08 01:39:24.917 [INFO][5421] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2", ResourceVersion:"935", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4", Pod:"coredns-668d6bf9bc-nsnp6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3403d4852c5", MAC:"c2:c2:db:4b:0f:bb", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:24.922607 containerd[1926]: 2025-11-08 01:39:24.921 [INFO][5421] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4" Namespace="kube-system" Pod="coredns-668d6bf9bc-nsnp6" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:24.931020 containerd[1926]: time="2025-11-08T01:39:24.930978128Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:24.931020 containerd[1926]: time="2025-11-08T01:39:24.931006873Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:24.931121 containerd[1926]: time="2025-11-08T01:39:24.931021486Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:24.931230 containerd[1926]: time="2025-11-08T01:39:24.931212147Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:25.017030 containerd[1926]: time="2025-11-08T01:39:25.016991995Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-nsnp6,Uid:2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2,Namespace:kube-system,Attempt:1,} returns sandbox id \"6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4\"" Nov 8 01:39:25.018477 containerd[1926]: time="2025-11-08T01:39:25.018456924Z" level=info msg="CreateContainer within sandbox \"6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 8 01:39:25.022874 containerd[1926]: time="2025-11-08T01:39:25.022819630Z" level=info msg="CreateContainer within sandbox \"6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d8ffb99b396a9ead9dd5ff6cefb98b59f9217d030e72a6d9e824593b55d28399\"" Nov 8 01:39:25.023082 containerd[1926]: time="2025-11-08T01:39:25.023072327Z" level=info msg="StartContainer for \"d8ffb99b396a9ead9dd5ff6cefb98b59f9217d030e72a6d9e824593b55d28399\"" Nov 8 01:39:25.024938 systemd-networkd[1553]: calib14bc7ba2cb: Link UP Nov 8 01:39:25.025086 systemd-networkd[1553]: calib14bc7ba2cb: Gained carrier Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:24.813 [INFO][5451] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0 coredns-668d6bf9bc- kube-system 4f0fc8a6-454f-454b-80b2-7ee71f1fc322 937 0 2025-11-08 01:38:51 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 coredns-668d6bf9bc-l4wcp eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calib14bc7ba2cb [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:24.813 [INFO][5451] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:24.826 [INFO][5501] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" HandleID="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:24.826 [INFO][5501] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" HandleID="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f670), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"coredns-668d6bf9bc-l4wcp", "timestamp":"2025-11-08 01:39:24.826287412 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:24.826 [INFO][5501] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:24.915 [INFO][5501] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:24.915 [INFO][5501] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.004 [INFO][5501] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.009 [INFO][5501] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.012 [INFO][5501] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.014 [INFO][5501] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.016 [INFO][5501] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.016 [INFO][5501] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.017 [INFO][5501] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.020 [INFO][5501] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.023 [INFO][5501] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.68/26] block=192.168.68.64/26 handle="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.023 [INFO][5501] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.68/26] handle="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.023 [INFO][5501] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:25.031540 containerd[1926]: 2025-11-08 01:39:25.023 [INFO][5501] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.68/26] IPv6=[] ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" HandleID="k8s-pod-network.beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:25.031934 containerd[1926]: 2025-11-08 01:39:25.024 [INFO][5451] cni-plugin/k8s.go 418: Populated endpoint ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"4f0fc8a6-454f-454b-80b2-7ee71f1fc322", ResourceVersion:"937", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"coredns-668d6bf9bc-l4wcp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib14bc7ba2cb", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:25.031934 containerd[1926]: 2025-11-08 01:39:25.024 [INFO][5451] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.68/32] ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:25.031934 containerd[1926]: 2025-11-08 01:39:25.024 [INFO][5451] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib14bc7ba2cb ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:25.031934 containerd[1926]: 2025-11-08 01:39:25.025 [INFO][5451] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:25.031934 containerd[1926]: 2025-11-08 01:39:25.025 [INFO][5451] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"4f0fc8a6-454f-454b-80b2-7ee71f1fc322", ResourceVersion:"937", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac", Pod:"coredns-668d6bf9bc-l4wcp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib14bc7ba2cb", MAC:"da:c8:7c:58:e7:f9", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:25.031934 containerd[1926]: 2025-11-08 01:39:25.030 [INFO][5451] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac" Namespace="kube-system" Pod="coredns-668d6bf9bc-l4wcp" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:25.040203 containerd[1926]: time="2025-11-08T01:39:25.040128661Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:25.040203 containerd[1926]: time="2025-11-08T01:39:25.040163065Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:25.040203 containerd[1926]: time="2025-11-08T01:39:25.040170103Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:25.040313 containerd[1926]: time="2025-11-08T01:39:25.040211121Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:25.054812 containerd[1926]: time="2025-11-08T01:39:25.054767642Z" level=info msg="StartContainer for \"d8ffb99b396a9ead9dd5ff6cefb98b59f9217d030e72a6d9e824593b55d28399\" returns successfully" Nov 8 01:39:25.066955 containerd[1926]: time="2025-11-08T01:39:25.066916408Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-l4wcp,Uid:4f0fc8a6-454f-454b-80b2-7ee71f1fc322,Namespace:kube-system,Attempt:1,} returns sandbox id \"beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac\"" Nov 8 01:39:25.068426 containerd[1926]: time="2025-11-08T01:39:25.068403755Z" level=info msg="CreateContainer within sandbox \"beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 8 01:39:25.072198 containerd[1926]: time="2025-11-08T01:39:25.072170975Z" level=info msg="CreateContainer within sandbox \"beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"fe297d35af5a2e89b917216a4a0431be6e9a128585ebf28a3ce7bca7810e1cd3\"" Nov 8 01:39:25.072481 containerd[1926]: time="2025-11-08T01:39:25.072434742Z" level=info msg="StartContainer for \"fe297d35af5a2e89b917216a4a0431be6e9a128585ebf28a3ce7bca7810e1cd3\"" Nov 8 01:39:25.107427 containerd[1926]: time="2025-11-08T01:39:25.107402464Z" level=info msg="StartContainer for \"fe297d35af5a2e89b917216a4a0431be6e9a128585ebf28a3ce7bca7810e1cd3\" returns successfully" Nov 8 01:39:25.124723 systemd-networkd[1553]: caliec314f3d499: Link UP Nov 8 01:39:25.125085 systemd-networkd[1553]: caliec314f3d499: Gained carrier Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:24.812 [INFO][5448] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0 calico-kube-controllers-566b9cf58b- calico-system f5773a4d-1a91-4ae3-82ea-bf0402f8aa01 936 0 2025-11-08 01:39:03 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:566b9cf58b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 calico-kube-controllers-566b9cf58b-whc7z eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] caliec314f3d499 [] [] }} ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:24.812 [INFO][5448] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:24.826 [INFO][5499] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" HandleID="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:24.826 [INFO][5499] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" HandleID="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000351150), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"calico-kube-controllers-566b9cf58b-whc7z", "timestamp":"2025-11-08 01:39:24.826576339 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:24.826 [INFO][5499] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.023 [INFO][5499] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.023 [INFO][5499] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.103 [INFO][5499] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.109 [INFO][5499] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.113 [INFO][5499] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.114 [INFO][5499] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.115 [INFO][5499] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.115 [INFO][5499] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.116 [INFO][5499] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4 Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.119 [INFO][5499] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.122 [INFO][5499] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.69/26] block=192.168.68.64/26 handle="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.122 [INFO][5499] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.69/26] handle="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.122 [INFO][5499] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:25.131728 containerd[1926]: 2025-11-08 01:39:25.122 [INFO][5499] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.69/26] IPv6=[] ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" HandleID="k8s-pod-network.6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:25.132275 containerd[1926]: 2025-11-08 01:39:25.123 [INFO][5448] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0", GenerateName:"calico-kube-controllers-566b9cf58b-", Namespace:"calico-system", SelfLink:"", UID:"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01", ResourceVersion:"936", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"566b9cf58b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"calico-kube-controllers-566b9cf58b-whc7z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"caliec314f3d499", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:25.132275 containerd[1926]: 2025-11-08 01:39:25.123 [INFO][5448] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.69/32] ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:25.132275 containerd[1926]: 2025-11-08 01:39:25.123 [INFO][5448] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliec314f3d499 ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:25.132275 containerd[1926]: 2025-11-08 01:39:25.124 [INFO][5448] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:25.132275 containerd[1926]: 2025-11-08 01:39:25.125 [INFO][5448] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0", GenerateName:"calico-kube-controllers-566b9cf58b-", Namespace:"calico-system", SelfLink:"", UID:"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01", ResourceVersion:"936", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"566b9cf58b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4", Pod:"calico-kube-controllers-566b9cf58b-whc7z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"caliec314f3d499", MAC:"5a:74:5b:d1:66:ad", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:25.132275 containerd[1926]: 2025-11-08 01:39:25.130 [INFO][5448] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4" Namespace="calico-system" Pod="calico-kube-controllers-566b9cf58b-whc7z" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:25.140329 containerd[1926]: time="2025-11-08T01:39:25.140253478Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:25.140502 containerd[1926]: time="2025-11-08T01:39:25.140452964Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:25.140502 containerd[1926]: time="2025-11-08T01:39:25.140462713Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:25.140558 containerd[1926]: time="2025-11-08T01:39:25.140507217Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:25.183382 containerd[1926]: time="2025-11-08T01:39:25.183332024Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-566b9cf58b-whc7z,Uid:f5773a4d-1a91-4ae3-82ea-bf0402f8aa01,Namespace:calico-system,Attempt:1,} returns sandbox id \"6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4\"" Nov 8 01:39:25.298756 containerd[1926]: time="2025-11-08T01:39:25.298658602Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:25.299599 containerd[1926]: time="2025-11-08T01:39:25.299550572Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:39:25.299636 containerd[1926]: time="2025-11-08T01:39:25.299597878Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:39:25.299741 kubelet[3256]: E1108 01:39:25.299688 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:25.299741 kubelet[3256]: E1108 01:39:25.299720 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:25.299971 kubelet[3256]: E1108 01:39:25.299862 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv77z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:25.300035 containerd[1926]: time="2025-11-08T01:39:25.299918464Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 8 01:39:25.300986 kubelet[3256]: E1108 01:39:25.300971 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:39:25.671208 containerd[1926]: time="2025-11-08T01:39:25.670937321Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:25.672011 containerd[1926]: time="2025-11-08T01:39:25.671927288Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 8 01:39:25.672049 containerd[1926]: time="2025-11-08T01:39:25.671995385Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 8 01:39:25.672093 kubelet[3256]: E1108 01:39:25.672073 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:39:25.672119 kubelet[3256]: E1108 01:39:25.672103 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:39:25.672295 kubelet[3256]: E1108 01:39:25.672229 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cts9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:25.673417 kubelet[3256]: E1108 01:39:25.673373 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:39:25.794470 systemd[1]: run-netns-cni\x2d2ac0e4cf\x2d3444\x2d456d\x2d2c80\x2d92512319f39d.mount: Deactivated successfully. Nov 8 01:39:25.828210 kubelet[3256]: E1108 01:39:25.828177 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:39:25.829100 kubelet[3256]: E1108 01:39:25.829073 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:39:25.835957 kubelet[3256]: I1108 01:39:25.835909 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-nsnp6" podStartSLOduration=34.835891183 podStartE2EDuration="34.835891183s" podCreationTimestamp="2025-11-08 01:38:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-08 01:39:25.835654203 +0000 UTC m=+40.167834967" watchObservedRunningTime="2025-11-08 01:39:25.835891183 +0000 UTC m=+40.168071941" Nov 8 01:39:25.853507 kubelet[3256]: I1108 01:39:25.853442 3256 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-l4wcp" podStartSLOduration=34.853421301 podStartE2EDuration="34.853421301s" podCreationTimestamp="2025-11-08 01:38:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-08 01:39:25.853288359 +0000 UTC m=+40.185469128" watchObservedRunningTime="2025-11-08 01:39:25.853421301 +0000 UTC m=+40.185602063" Nov 8 01:39:26.177391 systemd-networkd[1553]: calib14bc7ba2cb: Gained IPv6LL Nov 8 01:39:26.433472 systemd-networkd[1553]: caliec314f3d499: Gained IPv6LL Nov 8 01:39:26.626496 systemd-networkd[1553]: cali3403d4852c5: Gained IPv6LL Nov 8 01:39:26.753481 systemd-networkd[1553]: calife27e0b4819: Gained IPv6LL Nov 8 01:39:26.834876 kubelet[3256]: E1108 01:39:26.834795 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:39:26.836008 kubelet[3256]: E1108 01:39:26.835085 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:39:27.718008 containerd[1926]: time="2025-11-08T01:39:27.717892113Z" level=info msg="StopPodSandbox for \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\"" Nov 8 01:39:27.718008 containerd[1926]: time="2025-11-08T01:39:27.717976482Z" level=info msg="StopPodSandbox for \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\"" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5876] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5876] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" iface="eth0" netns="/var/run/netns/cni-aa977fa2-a49d-37f8-85be-aaecbc2b6615" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5876] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" iface="eth0" netns="/var/run/netns/cni-aa977fa2-a49d-37f8-85be-aaecbc2b6615" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5876] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" iface="eth0" netns="/var/run/netns/cni-aa977fa2-a49d-37f8-85be-aaecbc2b6615" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5876] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5876] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.823 [INFO][5914] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.823 [INFO][5914] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.823 [INFO][5914] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.831 [WARNING][5914] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.831 [INFO][5914] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.833 [INFO][5914] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:27.836153 containerd[1926]: 2025-11-08 01:39:27.834 [INFO][5876] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:27.836702 containerd[1926]: time="2025-11-08T01:39:27.836274107Z" level=info msg="TearDown network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\" successfully" Nov 8 01:39:27.836702 containerd[1926]: time="2025-11-08T01:39:27.836301443Z" level=info msg="StopPodSandbox for \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\" returns successfully" Nov 8 01:39:27.836864 containerd[1926]: time="2025-11-08T01:39:27.836808319Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dmqbg,Uid:ff4b01f8-4bf6-4050-afe4-e2ac586eece3,Namespace:calico-system,Attempt:1,}" Nov 8 01:39:27.839095 systemd[1]: run-netns-cni\x2daa977fa2\x2da49d\x2d37f8\x2d85be\x2daaecbc2b6615.mount: Deactivated successfully. Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.791 [INFO][5875] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5875] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" iface="eth0" netns="/var/run/netns/cni-6ae0171e-dc0e-c11b-3a38-13d3133f9db6" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5875] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" iface="eth0" netns="/var/run/netns/cni-6ae0171e-dc0e-c11b-3a38-13d3133f9db6" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5875] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" iface="eth0" netns="/var/run/netns/cni-6ae0171e-dc0e-c11b-3a38-13d3133f9db6" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5875] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.792 [INFO][5875] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.823 [INFO][5913] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.823 [INFO][5913] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.833 [INFO][5913] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.839 [WARNING][5913] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.839 [INFO][5913] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.840 [INFO][5913] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:27.843029 containerd[1926]: 2025-11-08 01:39:27.841 [INFO][5875] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:27.843304 containerd[1926]: time="2025-11-08T01:39:27.843115295Z" level=info msg="TearDown network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\" successfully" Nov 8 01:39:27.843304 containerd[1926]: time="2025-11-08T01:39:27.843131618Z" level=info msg="StopPodSandbox for \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\" returns successfully" Nov 8 01:39:27.843547 containerd[1926]: time="2025-11-08T01:39:27.843535852Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-gztt5,Uid:31ac439b-7263-4e83-932f-3464165d69ab,Namespace:calico-apiserver,Attempt:1,}" Nov 8 01:39:27.845327 systemd[1]: run-netns-cni\x2d6ae0171e\x2ddc0e\x2dc11b\x2d3a38\x2d13d3133f9db6.mount: Deactivated successfully. Nov 8 01:39:27.889633 systemd-networkd[1553]: cali345f126179d: Link UP Nov 8 01:39:27.889776 systemd-networkd[1553]: cali345f126179d: Gained carrier Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.858 [INFO][5944] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0 csi-node-driver- calico-system ff4b01f8-4bf6-4050-afe4-e2ac586eece3 1002 0 2025-11-08 01:39:03 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 csi-node-driver-dmqbg eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali345f126179d [] [] }} ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.858 [INFO][5944] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.871 [INFO][5990] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" HandleID="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.871 [INFO][5990] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" HandleID="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000137690), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"csi-node-driver-dmqbg", "timestamp":"2025-11-08 01:39:27.871298133 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.871 [INFO][5990] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.871 [INFO][5990] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.871 [INFO][5990] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.875 [INFO][5990] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.877 [INFO][5990] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.879 [INFO][5990] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.880 [INFO][5990] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.882 [INFO][5990] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.882 [INFO][5990] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.882 [INFO][5990] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344 Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.884 [INFO][5990] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.887 [INFO][5990] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.70/26] block=192.168.68.64/26 handle="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.887 [INFO][5990] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.70/26] handle="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.887 [INFO][5990] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:27.895532 containerd[1926]: 2025-11-08 01:39:27.887 [INFO][5990] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.70/26] IPv6=[] ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" HandleID="k8s-pod-network.7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.896032 containerd[1926]: 2025-11-08 01:39:27.888 [INFO][5944] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ff4b01f8-4bf6-4050-afe4-e2ac586eece3", ResourceVersion:"1002", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"csi-node-driver-dmqbg", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali345f126179d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:27.896032 containerd[1926]: 2025-11-08 01:39:27.888 [INFO][5944] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.70/32] ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.896032 containerd[1926]: 2025-11-08 01:39:27.888 [INFO][5944] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali345f126179d ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.896032 containerd[1926]: 2025-11-08 01:39:27.889 [INFO][5944] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.896032 containerd[1926]: 2025-11-08 01:39:27.890 [INFO][5944] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ff4b01f8-4bf6-4050-afe4-e2ac586eece3", ResourceVersion:"1002", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344", Pod:"csi-node-driver-dmqbg", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali345f126179d", MAC:"26:dc:ab:38:51:eb", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:27.896032 containerd[1926]: 2025-11-08 01:39:27.894 [INFO][5944] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344" Namespace="calico-system" Pod="csi-node-driver-dmqbg" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:27.903909 containerd[1926]: time="2025-11-08T01:39:27.903692108Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:27.903909 containerd[1926]: time="2025-11-08T01:39:27.903903362Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:27.903909 containerd[1926]: time="2025-11-08T01:39:27.903911801Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:27.904010 containerd[1926]: time="2025-11-08T01:39:27.903970276Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:27.924692 containerd[1926]: time="2025-11-08T01:39:27.924643613Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dmqbg,Uid:ff4b01f8-4bf6-4050-afe4-e2ac586eece3,Namespace:calico-system,Attempt:1,} returns sandbox id \"7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344\"" Nov 8 01:39:27.925363 containerd[1926]: time="2025-11-08T01:39:27.925348408Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 8 01:39:28.010425 systemd-networkd[1553]: cali3d27f329b87: Link UP Nov 8 01:39:28.010663 systemd-networkd[1553]: cali3d27f329b87: Gained carrier Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.863 [INFO][5961] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0 calico-apiserver-66c4f75c9f- calico-apiserver 31ac439b-7263-4e83-932f-3464165d69ab 1001 0 2025-11-08 01:38:58 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:66c4f75c9f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 calico-apiserver-66c4f75c9f-gztt5 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali3d27f329b87 [] [] }} ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.863 [INFO][5961] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.875 [INFO][5996] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" HandleID="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.875 [INFO][5996] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" HandleID="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000139ec0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"calico-apiserver-66c4f75c9f-gztt5", "timestamp":"2025-11-08 01:39:27.875397536 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.875 [INFO][5996] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.887 [INFO][5996] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.887 [INFO][5996] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.976 [INFO][5996] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.983 [INFO][5996] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.989 [INFO][5996] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.992 [INFO][5996] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.997 [INFO][5996] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:27.997 [INFO][5996] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:28.000 [INFO][5996] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3 Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:28.004 [INFO][5996] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:28.007 [INFO][5996] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.71/26] block=192.168.68.64/26 handle="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:28.007 [INFO][5996] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.71/26] handle="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:28.007 [INFO][5996] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:28.042291 containerd[1926]: 2025-11-08 01:39:28.007 [INFO][5996] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.71/26] IPv6=[] ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" HandleID="k8s-pod-network.3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:28.044398 containerd[1926]: 2025-11-08 01:39:28.009 [INFO][5961] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"31ac439b-7263-4e83-932f-3464165d69ab", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"calico-apiserver-66c4f75c9f-gztt5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3d27f329b87", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:28.044398 containerd[1926]: 2025-11-08 01:39:28.009 [INFO][5961] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.71/32] ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:28.044398 containerd[1926]: 2025-11-08 01:39:28.009 [INFO][5961] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3d27f329b87 ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:28.044398 containerd[1926]: 2025-11-08 01:39:28.010 [INFO][5961] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:28.044398 containerd[1926]: 2025-11-08 01:39:28.011 [INFO][5961] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"31ac439b-7263-4e83-932f-3464165d69ab", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3", Pod:"calico-apiserver-66c4f75c9f-gztt5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3d27f329b87", MAC:"ba:93:88:b5:9e:85", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:28.044398 containerd[1926]: 2025-11-08 01:39:28.037 [INFO][5961] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3" Namespace="calico-apiserver" Pod="calico-apiserver-66c4f75c9f-gztt5" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:28.056275 containerd[1926]: time="2025-11-08T01:39:28.056108547Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:28.056275 containerd[1926]: time="2025-11-08T01:39:28.056267387Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:28.056275 containerd[1926]: time="2025-11-08T01:39:28.056277088Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:28.056381 containerd[1926]: time="2025-11-08T01:39:28.056342172Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:28.134653 containerd[1926]: time="2025-11-08T01:39:28.134621390Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-66c4f75c9f-gztt5,Uid:31ac439b-7263-4e83-932f-3464165d69ab,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3\"" Nov 8 01:39:28.352622 containerd[1926]: time="2025-11-08T01:39:28.352466679Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:28.353499 containerd[1926]: time="2025-11-08T01:39:28.353409138Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 8 01:39:28.353499 containerd[1926]: time="2025-11-08T01:39:28.353479803Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 8 01:39:28.353746 kubelet[3256]: E1108 01:39:28.353675 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:39:28.353746 kubelet[3256]: E1108 01:39:28.353711 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:39:28.354063 kubelet[3256]: E1108 01:39:28.353846 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:28.354121 containerd[1926]: time="2025-11-08T01:39:28.353984251Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:39:28.713238 containerd[1926]: time="2025-11-08T01:39:28.712974672Z" level=info msg="StopPodSandbox for \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\"" Nov 8 01:39:28.723904 containerd[1926]: time="2025-11-08T01:39:28.723778346Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:28.724711 containerd[1926]: time="2025-11-08T01:39:28.724642224Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:39:28.724753 containerd[1926]: time="2025-11-08T01:39:28.724705216Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:39:28.724841 kubelet[3256]: E1108 01:39:28.724797 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:28.724841 kubelet[3256]: E1108 01:39:28.724824 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:28.724996 kubelet[3256]: E1108 01:39:28.724950 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b49rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:28.725114 containerd[1926]: time="2025-11-08T01:39:28.725102058Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 8 01:39:28.726221 kubelet[3256]: E1108 01:39:28.726205 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.744 [INFO][6135] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.744 [INFO][6135] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" iface="eth0" netns="/var/run/netns/cni-ceffab8f-5200-7375-65f7-cb376bf7f315" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.744 [INFO][6135] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" iface="eth0" netns="/var/run/netns/cni-ceffab8f-5200-7375-65f7-cb376bf7f315" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.744 [INFO][6135] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" iface="eth0" netns="/var/run/netns/cni-ceffab8f-5200-7375-65f7-cb376bf7f315" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.744 [INFO][6135] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.744 [INFO][6135] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.755 [INFO][6152] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.755 [INFO][6152] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.755 [INFO][6152] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.759 [WARNING][6152] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.759 [INFO][6152] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.760 [INFO][6152] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:28.762271 containerd[1926]: 2025-11-08 01:39:28.761 [INFO][6135] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:28.762614 containerd[1926]: time="2025-11-08T01:39:28.762330158Z" level=info msg="TearDown network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\" successfully" Nov 8 01:39:28.762614 containerd[1926]: time="2025-11-08T01:39:28.762348256Z" level=info msg="StopPodSandbox for \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\" returns successfully" Nov 8 01:39:28.762748 containerd[1926]: time="2025-11-08T01:39:28.762733384Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-8ld4r,Uid:5f897823-41ac-4b66-8214-3b03dd8d6290,Namespace:calico-system,Attempt:1,}" Nov 8 01:39:28.812615 systemd-networkd[1553]: cali5477facc711: Link UP Nov 8 01:39:28.812786 systemd-networkd[1553]: cali5477facc711: Gained carrier Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.782 [INFO][6168] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0 goldmane-666569f655- calico-system 5f897823-41ac-4b66-8214-3b03dd8d6290 1018 0 2025-11-08 01:39:01 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081.3.6-n-3863d1ecf5 goldmane-666569f655-8ld4r eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali5477facc711 [] [] }} ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.783 [INFO][6168] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.794 [INFO][6188] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" HandleID="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.794 [INFO][6188] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" HandleID="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f6b0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-3863d1ecf5", "pod":"goldmane-666569f655-8ld4r", "timestamp":"2025-11-08 01:39:28.79479267 +0000 UTC"}, Hostname:"ci-4081.3.6-n-3863d1ecf5", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.794 [INFO][6188] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.794 [INFO][6188] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.794 [INFO][6188] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-3863d1ecf5' Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.798 [INFO][6188] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.800 [INFO][6188] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.802 [INFO][6188] ipam/ipam.go 511: Trying affinity for 192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.803 [INFO][6188] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.805 [INFO][6188] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.64/26 host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.805 [INFO][6188] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.68.64/26 handle="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.805 [INFO][6188] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2 Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.807 [INFO][6188] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.68.64/26 handle="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.810 [INFO][6188] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.68.72/26] block=192.168.68.64/26 handle="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.810 [INFO][6188] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.72/26] handle="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" host="ci-4081.3.6-n-3863d1ecf5" Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.810 [INFO][6188] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:28.818933 containerd[1926]: 2025-11-08 01:39:28.810 [INFO][6188] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.68.72/26] IPv6=[] ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" HandleID="k8s-pod-network.eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.819448 containerd[1926]: 2025-11-08 01:39:28.811 [INFO][6168] cni-plugin/k8s.go 418: Populated endpoint ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"5f897823-41ac-4b66-8214-3b03dd8d6290", ResourceVersion:"1018", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"", Pod:"goldmane-666569f655-8ld4r", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali5477facc711", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:28.819448 containerd[1926]: 2025-11-08 01:39:28.811 [INFO][6168] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.72/32] ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.819448 containerd[1926]: 2025-11-08 01:39:28.811 [INFO][6168] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5477facc711 ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.819448 containerd[1926]: 2025-11-08 01:39:28.812 [INFO][6168] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.819448 containerd[1926]: 2025-11-08 01:39:28.812 [INFO][6168] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"5f897823-41ac-4b66-8214-3b03dd8d6290", ResourceVersion:"1018", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2", Pod:"goldmane-666569f655-8ld4r", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali5477facc711", MAC:"96:76:27:4e:4f:a9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:28.819448 containerd[1926]: 2025-11-08 01:39:28.818 [INFO][6168] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2" Namespace="calico-system" Pod="goldmane-666569f655-8ld4r" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:28.827649 containerd[1926]: time="2025-11-08T01:39:28.827601463Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 8 01:39:28.827649 containerd[1926]: time="2025-11-08T01:39:28.827631029Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 8 01:39:28.827649 containerd[1926]: time="2025-11-08T01:39:28.827638268Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:28.827776 containerd[1926]: time="2025-11-08T01:39:28.827682299Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 8 01:39:28.836716 kubelet[3256]: E1108 01:39:28.836697 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:39:28.840766 systemd[1]: run-netns-cni\x2dceffab8f\x2d5200\x2d7375\x2d65f7\x2dcb376bf7f315.mount: Deactivated successfully. Nov 8 01:39:28.865453 containerd[1926]: time="2025-11-08T01:39:28.865433915Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-8ld4r,Uid:5f897823-41ac-4b66-8214-3b03dd8d6290,Namespace:calico-system,Attempt:1,} returns sandbox id \"eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2\"" Nov 8 01:39:29.089384 containerd[1926]: time="2025-11-08T01:39:29.089273551Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:29.090388 containerd[1926]: time="2025-11-08T01:39:29.090361022Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 8 01:39:29.090446 containerd[1926]: time="2025-11-08T01:39:29.090419220Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 8 01:39:29.090589 kubelet[3256]: E1108 01:39:29.090539 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:39:29.090589 kubelet[3256]: E1108 01:39:29.090572 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:39:29.090757 kubelet[3256]: E1108 01:39:29.090706 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:29.090827 containerd[1926]: time="2025-11-08T01:39:29.090779338Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 8 01:39:29.091868 kubelet[3256]: E1108 01:39:29.091826 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:29.377631 systemd-networkd[1553]: cali345f126179d: Gained IPv6LL Nov 8 01:39:29.441473 systemd-networkd[1553]: cali3d27f329b87: Gained IPv6LL Nov 8 01:39:29.467764 containerd[1926]: time="2025-11-08T01:39:29.467643700Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:29.468502 containerd[1926]: time="2025-11-08T01:39:29.468427273Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 8 01:39:29.468502 containerd[1926]: time="2025-11-08T01:39:29.468490877Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 8 01:39:29.468660 kubelet[3256]: E1108 01:39:29.468609 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:39:29.468660 kubelet[3256]: E1108 01:39:29.468639 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:39:29.468863 kubelet[3256]: E1108 01:39:29.468728 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rplv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:29.469885 kubelet[3256]: E1108 01:39:29.469841 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:39:29.845471 kubelet[3256]: E1108 01:39:29.845387 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:39:29.845803 kubelet[3256]: E1108 01:39:29.845460 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:39:29.846359 kubelet[3256]: E1108 01:39:29.846234 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:30.529508 systemd-networkd[1553]: cali5477facc711: Gained IPv6LL Nov 8 01:39:30.848954 kubelet[3256]: E1108 01:39:30.848728 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:39:34.715358 containerd[1926]: time="2025-11-08T01:39:34.715263818Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 8 01:39:35.082781 containerd[1926]: time="2025-11-08T01:39:35.082679043Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:35.083718 containerd[1926]: time="2025-11-08T01:39:35.083629109Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 8 01:39:35.083718 containerd[1926]: time="2025-11-08T01:39:35.083690850Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 8 01:39:35.083809 kubelet[3256]: E1108 01:39:35.083783 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:39:35.084008 kubelet[3256]: E1108 01:39:35.083818 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:39:35.084008 kubelet[3256]: E1108 01:39:35.083884 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:52eaa7a533b1456aa7b1871d4874688d,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:35.085620 containerd[1926]: time="2025-11-08T01:39:35.085573043Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 8 01:39:35.472713 containerd[1926]: time="2025-11-08T01:39:35.472488633Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:35.473313 containerd[1926]: time="2025-11-08T01:39:35.473296814Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 8 01:39:35.473359 containerd[1926]: time="2025-11-08T01:39:35.473339824Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 8 01:39:35.473459 kubelet[3256]: E1108 01:39:35.473436 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:39:35.473500 kubelet[3256]: E1108 01:39:35.473468 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:39:35.473549 kubelet[3256]: E1108 01:39:35.473531 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:35.474703 kubelet[3256]: E1108 01:39:35.474687 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:39:37.711888 containerd[1926]: time="2025-11-08T01:39:37.711868704Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 8 01:39:38.087128 containerd[1926]: time="2025-11-08T01:39:38.087031362Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:38.088140 containerd[1926]: time="2025-11-08T01:39:38.088113325Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 8 01:39:38.088203 containerd[1926]: time="2025-11-08T01:39:38.088186349Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 8 01:39:38.088295 kubelet[3256]: E1108 01:39:38.088272 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:39:38.088485 kubelet[3256]: E1108 01:39:38.088305 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:39:38.088485 kubelet[3256]: E1108 01:39:38.088381 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cts9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:38.089566 kubelet[3256]: E1108 01:39:38.089553 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:39:40.712095 containerd[1926]: time="2025-11-08T01:39:40.712072895Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:39:41.103129 containerd[1926]: time="2025-11-08T01:39:41.102995365Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:41.103978 containerd[1926]: time="2025-11-08T01:39:41.103886543Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:39:41.104011 containerd[1926]: time="2025-11-08T01:39:41.103974925Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:39:41.104151 kubelet[3256]: E1108 01:39:41.104125 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:41.104483 kubelet[3256]: E1108 01:39:41.104162 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:41.104483 kubelet[3256]: E1108 01:39:41.104360 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv77z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:41.104616 containerd[1926]: time="2025-11-08T01:39:41.104473069Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:39:41.105583 kubelet[3256]: E1108 01:39:41.105539 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:39:41.479051 containerd[1926]: time="2025-11-08T01:39:41.478773094Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:41.479928 containerd[1926]: time="2025-11-08T01:39:41.479835597Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:39:41.479990 containerd[1926]: time="2025-11-08T01:39:41.479925430Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:39:41.480087 kubelet[3256]: E1108 01:39:41.480067 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:41.480113 kubelet[3256]: E1108 01:39:41.480096 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:39:41.480283 kubelet[3256]: E1108 01:39:41.480193 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b49rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:41.481371 kubelet[3256]: E1108 01:39:41.481327 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:39:41.713898 containerd[1926]: time="2025-11-08T01:39:41.713824962Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 8 01:39:42.091084 containerd[1926]: time="2025-11-08T01:39:42.090953844Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:42.091963 containerd[1926]: time="2025-11-08T01:39:42.091867686Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 8 01:39:42.091963 containerd[1926]: time="2025-11-08T01:39:42.091933570Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 8 01:39:42.092237 kubelet[3256]: E1108 01:39:42.092168 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:39:42.092237 kubelet[3256]: E1108 01:39:42.092217 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:39:42.092389 kubelet[3256]: E1108 01:39:42.092311 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:42.094145 containerd[1926]: time="2025-11-08T01:39:42.094082429Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 8 01:39:42.464677 containerd[1926]: time="2025-11-08T01:39:42.464411533Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:42.465466 containerd[1926]: time="2025-11-08T01:39:42.465387412Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 8 01:39:42.465548 containerd[1926]: time="2025-11-08T01:39:42.465454243Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 8 01:39:42.465707 kubelet[3256]: E1108 01:39:42.465634 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:39:42.465707 kubelet[3256]: E1108 01:39:42.465683 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:39:42.465973 kubelet[3256]: E1108 01:39:42.465767 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:42.466998 kubelet[3256]: E1108 01:39:42.466953 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:43.712022 containerd[1926]: time="2025-11-08T01:39:43.712000585Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 8 01:39:44.085025 containerd[1926]: time="2025-11-08T01:39:44.084885438Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:39:44.086056 containerd[1926]: time="2025-11-08T01:39:44.085967267Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 8 01:39:44.086117 containerd[1926]: time="2025-11-08T01:39:44.086053695Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 8 01:39:44.086267 kubelet[3256]: E1108 01:39:44.086199 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:39:44.086267 kubelet[3256]: E1108 01:39:44.086262 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:39:44.086554 kubelet[3256]: E1108 01:39:44.086351 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rplv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 8 01:39:44.087554 kubelet[3256]: E1108 01:39:44.087511 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:39:45.707216 containerd[1926]: time="2025-11-08T01:39:45.707195250Z" level=info msg="StopPodSandbox for \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\"" Nov 8 01:39:45.711757 kubelet[3256]: E1108 01:39:45.711730 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.727 [WARNING][6284] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"31ac439b-7263-4e83-932f-3464165d69ab", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3", Pod:"calico-apiserver-66c4f75c9f-gztt5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3d27f329b87", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.727 [INFO][6284] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.727 [INFO][6284] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" iface="eth0" netns="" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.727 [INFO][6284] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.727 [INFO][6284] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.740 [INFO][6303] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.740 [INFO][6303] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.740 [INFO][6303] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.744 [WARNING][6303] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.744 [INFO][6303] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.745 [INFO][6303] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:45.746819 containerd[1926]: 2025-11-08 01:39:45.746 [INFO][6284] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.746819 containerd[1926]: time="2025-11-08T01:39:45.746813760Z" level=info msg="TearDown network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\" successfully" Nov 8 01:39:45.747184 containerd[1926]: time="2025-11-08T01:39:45.746831963Z" level=info msg="StopPodSandbox for \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\" returns successfully" Nov 8 01:39:45.747184 containerd[1926]: time="2025-11-08T01:39:45.747132326Z" level=info msg="RemovePodSandbox for \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\"" Nov 8 01:39:45.747184 containerd[1926]: time="2025-11-08T01:39:45.747155343Z" level=info msg="Forcibly stopping sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\"" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.764 [WARNING][6325] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"31ac439b-7263-4e83-932f-3464165d69ab", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"3dfe2bf54e77a7b3e00703147a3ab2636244703e75ba200efe6cf2f6d8dbc7c3", Pod:"calico-apiserver-66c4f75c9f-gztt5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3d27f329b87", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.764 [INFO][6325] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.764 [INFO][6325] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" iface="eth0" netns="" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.764 [INFO][6325] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.764 [INFO][6325] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.775 [INFO][6343] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.775 [INFO][6343] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.775 [INFO][6343] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.779 [WARNING][6343] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.779 [INFO][6343] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" HandleID="k8s-pod-network.e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--gztt5-eth0" Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.780 [INFO][6343] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:45.781966 containerd[1926]: 2025-11-08 01:39:45.781 [INFO][6325] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c" Nov 8 01:39:45.782324 containerd[1926]: time="2025-11-08T01:39:45.781992706Z" level=info msg="TearDown network for sandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\" successfully" Nov 8 01:39:45.783533 containerd[1926]: time="2025-11-08T01:39:45.783519826Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:45.783585 containerd[1926]: time="2025-11-08T01:39:45.783549716Z" level=info msg="RemovePodSandbox \"e8d85b7f2e7b4e8ba1c57dcc66111a938dfc7245117dce7e64614f96a9d7957c\" returns successfully" Nov 8 01:39:45.783844 containerd[1926]: time="2025-11-08T01:39:45.783833211Z" level=info msg="StopPodSandbox for \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\"" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.800 [WARNING][6369] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2", ResourceVersion:"980", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4", Pod:"coredns-668d6bf9bc-nsnp6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3403d4852c5", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.800 [INFO][6369] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.800 [INFO][6369] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" iface="eth0" netns="" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.800 [INFO][6369] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.800 [INFO][6369] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.812 [INFO][6385] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.812 [INFO][6385] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.812 [INFO][6385] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.816 [WARNING][6385] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.816 [INFO][6385] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.817 [INFO][6385] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:45.819552 containerd[1926]: 2025-11-08 01:39:45.818 [INFO][6369] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.819953 containerd[1926]: time="2025-11-08T01:39:45.819551882Z" level=info msg="TearDown network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\" successfully" Nov 8 01:39:45.819953 containerd[1926]: time="2025-11-08T01:39:45.819570221Z" level=info msg="StopPodSandbox for \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\" returns successfully" Nov 8 01:39:45.819953 containerd[1926]: time="2025-11-08T01:39:45.819899076Z" level=info msg="RemovePodSandbox for \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\"" Nov 8 01:39:45.819953 containerd[1926]: time="2025-11-08T01:39:45.819920758Z" level=info msg="Forcibly stopping sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\"" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.840 [WARNING][6411] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"2f5e0c1f-c9d6-4271-89ff-f7c97fc99df2", ResourceVersion:"980", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"6a994688466810d2d39c4d9bf2ddff4a952ac183ba51a125a99a8243bb50fae4", Pod:"coredns-668d6bf9bc-nsnp6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3403d4852c5", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.840 [INFO][6411] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.840 [INFO][6411] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" iface="eth0" netns="" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.840 [INFO][6411] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.840 [INFO][6411] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.853 [INFO][6428] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.853 [INFO][6428] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.853 [INFO][6428] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.858 [WARNING][6428] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.858 [INFO][6428] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" HandleID="k8s-pod-network.4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--nsnp6-eth0" Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.859 [INFO][6428] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:45.861435 containerd[1926]: 2025-11-08 01:39:45.860 [INFO][6411] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900" Nov 8 01:39:45.861811 containerd[1926]: time="2025-11-08T01:39:45.861440353Z" level=info msg="TearDown network for sandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\" successfully" Nov 8 01:39:45.863022 containerd[1926]: time="2025-11-08T01:39:45.862980371Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:45.863022 containerd[1926]: time="2025-11-08T01:39:45.863011497Z" level=info msg="RemovePodSandbox \"4e0b4d9633e1c03921f97fe25088c816eea7d21ad3cb94468c7b93595e48c900\" returns successfully" Nov 8 01:39:45.863265 containerd[1926]: time="2025-11-08T01:39:45.863236273Z" level=info msg="StopPodSandbox for \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\"" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.882 [WARNING][6454] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"4f0fc8a6-454f-454b-80b2-7ee71f1fc322", ResourceVersion:"974", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac", Pod:"coredns-668d6bf9bc-l4wcp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib14bc7ba2cb", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.882 [INFO][6454] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.883 [INFO][6454] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" iface="eth0" netns="" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.883 [INFO][6454] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.883 [INFO][6454] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.892 [INFO][6470] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.892 [INFO][6470] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.892 [INFO][6470] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.896 [WARNING][6470] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.896 [INFO][6470] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.897 [INFO][6470] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:45.899043 containerd[1926]: 2025-11-08 01:39:45.898 [INFO][6454] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.899345 containerd[1926]: time="2025-11-08T01:39:45.899067153Z" level=info msg="TearDown network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\" successfully" Nov 8 01:39:45.899345 containerd[1926]: time="2025-11-08T01:39:45.899082007Z" level=info msg="StopPodSandbox for \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\" returns successfully" Nov 8 01:39:45.899380 containerd[1926]: time="2025-11-08T01:39:45.899346881Z" level=info msg="RemovePodSandbox for \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\"" Nov 8 01:39:45.899380 containerd[1926]: time="2025-11-08T01:39:45.899364479Z" level=info msg="Forcibly stopping sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\"" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.915 [WARNING][6495] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"4f0fc8a6-454f-454b-80b2-7ee71f1fc322", ResourceVersion:"974", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"beea2d2ef517a384b7ecfa5260ba809d9684452be5cc33e07859543f1b5625ac", Pod:"coredns-668d6bf9bc-l4wcp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calib14bc7ba2cb", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.915 [INFO][6495] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.915 [INFO][6495] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" iface="eth0" netns="" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.916 [INFO][6495] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.916 [INFO][6495] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.925 [INFO][6514] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.925 [INFO][6514] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.925 [INFO][6514] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.929 [WARNING][6514] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.929 [INFO][6514] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" HandleID="k8s-pod-network.1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-coredns--668d6bf9bc--l4wcp-eth0" Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.930 [INFO][6514] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:45.932427 containerd[1926]: 2025-11-08 01:39:45.931 [INFO][6495] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6" Nov 8 01:39:45.932738 containerd[1926]: time="2025-11-08T01:39:45.932434624Z" level=info msg="TearDown network for sandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\" successfully" Nov 8 01:39:45.933791 containerd[1926]: time="2025-11-08T01:39:45.933748526Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:45.933791 containerd[1926]: time="2025-11-08T01:39:45.933773750Z" level=info msg="RemovePodSandbox \"1bdf8e4518eac851209aa698905a6f34abfe26e0fe2e713253210f8ef1df93a6\" returns successfully" Nov 8 01:39:45.934064 containerd[1926]: time="2025-11-08T01:39:45.934027549Z" level=info msg="StopPodSandbox for \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\"" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.952 [WARNING][6539] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"5f897823-41ac-4b66-8214-3b03dd8d6290", ResourceVersion:"1050", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2", Pod:"goldmane-666569f655-8ld4r", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali5477facc711", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.952 [INFO][6539] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.952 [INFO][6539] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" iface="eth0" netns="" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.952 [INFO][6539] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.952 [INFO][6539] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.964 [INFO][6555] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.964 [INFO][6555] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.964 [INFO][6555] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.969 [WARNING][6555] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.969 [INFO][6555] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.970 [INFO][6555] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:45.971971 containerd[1926]: 2025-11-08 01:39:45.971 [INFO][6539] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:45.971971 containerd[1926]: time="2025-11-08T01:39:45.971913363Z" level=info msg="TearDown network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\" successfully" Nov 8 01:39:45.971971 containerd[1926]: time="2025-11-08T01:39:45.971931131Z" level=info msg="StopPodSandbox for \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\" returns successfully" Nov 8 01:39:45.972364 containerd[1926]: time="2025-11-08T01:39:45.972221385Z" level=info msg="RemovePodSandbox for \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\"" Nov 8 01:39:45.972364 containerd[1926]: time="2025-11-08T01:39:45.972244335Z" level=info msg="Forcibly stopping sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\"" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:45.993 [WARNING][6580] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"5f897823-41ac-4b66-8214-3b03dd8d6290", ResourceVersion:"1050", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"eff4d4206878d49c3c2ee1d0df6923662eade054476e2ff2ad68daafef404ee2", Pod:"goldmane-666569f655-8ld4r", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali5477facc711", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:45.993 [INFO][6580] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:45.993 [INFO][6580] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" iface="eth0" netns="" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:45.993 [INFO][6580] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:45.993 [INFO][6580] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:46.005 [INFO][6597] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:46.005 [INFO][6597] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:46.005 [INFO][6597] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:46.011 [WARNING][6597] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:46.011 [INFO][6597] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" HandleID="k8s-pod-network.38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-goldmane--666569f655--8ld4r-eth0" Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:46.012 [INFO][6597] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.013989 containerd[1926]: 2025-11-08 01:39:46.013 [INFO][6580] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34" Nov 8 01:39:46.014597 containerd[1926]: time="2025-11-08T01:39:46.014022298Z" level=info msg="TearDown network for sandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\" successfully" Nov 8 01:39:46.015973 containerd[1926]: time="2025-11-08T01:39:46.015958397Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:46.016020 containerd[1926]: time="2025-11-08T01:39:46.015988351Z" level=info msg="RemovePodSandbox \"38ac13ff36bc00a3a6f6101a802554bea883dc785cc461224f63f0b24ba9ed34\" returns successfully" Nov 8 01:39:46.016248 containerd[1926]: time="2025-11-08T01:39:46.016234789Z" level=info msg="StopPodSandbox for \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\"" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.033 [WARNING][6621] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ff4b01f8-4bf6-4050-afe4-e2ac586eece3", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344", Pod:"csi-node-driver-dmqbg", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali345f126179d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.033 [INFO][6621] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.033 [INFO][6621] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" iface="eth0" netns="" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.033 [INFO][6621] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.034 [INFO][6621] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.044 [INFO][6637] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.044 [INFO][6637] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.044 [INFO][6637] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.048 [WARNING][6637] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.048 [INFO][6637] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.049 [INFO][6637] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.051287 containerd[1926]: 2025-11-08 01:39:46.050 [INFO][6621] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.051287 containerd[1926]: time="2025-11-08T01:39:46.051282359Z" level=info msg="TearDown network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\" successfully" Nov 8 01:39:46.051634 containerd[1926]: time="2025-11-08T01:39:46.051299692Z" level=info msg="StopPodSandbox for \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\" returns successfully" Nov 8 01:39:46.051634 containerd[1926]: time="2025-11-08T01:39:46.051573734Z" level=info msg="RemovePodSandbox for \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\"" Nov 8 01:39:46.051634 containerd[1926]: time="2025-11-08T01:39:46.051594447Z" level=info msg="Forcibly stopping sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\"" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.072 [WARNING][6663] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ff4b01f8-4bf6-4050-afe4-e2ac586eece3", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"7ede9618e1115adad80db20e0f8ffe743d02451ac44a52e4a151505f76f40344", Pod:"csi-node-driver-dmqbg", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali345f126179d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.072 [INFO][6663] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.072 [INFO][6663] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" iface="eth0" netns="" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.072 [INFO][6663] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.072 [INFO][6663] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.084 [INFO][6680] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.084 [INFO][6680] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.084 [INFO][6680] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.089 [WARNING][6680] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.089 [INFO][6680] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" HandleID="k8s-pod-network.a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-csi--node--driver--dmqbg-eth0" Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.090 [INFO][6680] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.091993 containerd[1926]: 2025-11-08 01:39:46.091 [INFO][6663] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11" Nov 8 01:39:46.092349 containerd[1926]: time="2025-11-08T01:39:46.092028595Z" level=info msg="TearDown network for sandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\" successfully" Nov 8 01:39:46.093579 containerd[1926]: time="2025-11-08T01:39:46.093539361Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:46.093579 containerd[1926]: time="2025-11-08T01:39:46.093565183Z" level=info msg="RemovePodSandbox \"a411b65c9a3a59dc2a40697719d611e9ee3813ac7d0a93dacab71f93a5c09f11\" returns successfully" Nov 8 01:39:46.093863 containerd[1926]: time="2025-11-08T01:39:46.093850388Z" level=info msg="StopPodSandbox for \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\"" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.110 [WARNING][6703] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"79b0696d-13d8-44db-9d91-6f1131e9b5ef", ResourceVersion:"993", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5", Pod:"calico-apiserver-66c4f75c9f-qhkcw", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calife27e0b4819", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.111 [INFO][6703] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.111 [INFO][6703] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" iface="eth0" netns="" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.111 [INFO][6703] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.111 [INFO][6703] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.121 [INFO][6719] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.121 [INFO][6719] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.121 [INFO][6719] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.125 [WARNING][6719] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.125 [INFO][6719] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.126 [INFO][6719] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.127730 containerd[1926]: 2025-11-08 01:39:46.127 [INFO][6703] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.128028 containerd[1926]: time="2025-11-08T01:39:46.127754136Z" level=info msg="TearDown network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\" successfully" Nov 8 01:39:46.128028 containerd[1926]: time="2025-11-08T01:39:46.127770547Z" level=info msg="StopPodSandbox for \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\" returns successfully" Nov 8 01:39:46.128061 containerd[1926]: time="2025-11-08T01:39:46.128045159Z" level=info msg="RemovePodSandbox for \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\"" Nov 8 01:39:46.128082 containerd[1926]: time="2025-11-08T01:39:46.128060256Z" level=info msg="Forcibly stopping sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\"" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.145 [WARNING][6746] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0", GenerateName:"calico-apiserver-66c4f75c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"79b0696d-13d8-44db-9d91-6f1131e9b5ef", ResourceVersion:"993", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 38, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"66c4f75c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"1200276f1ad3f442c722dce62cc4044c889b0648785c5742283ee683884269b5", Pod:"calico-apiserver-66c4f75c9f-qhkcw", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calife27e0b4819", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.145 [INFO][6746] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.145 [INFO][6746] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" iface="eth0" netns="" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.145 [INFO][6746] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.145 [INFO][6746] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.155 [INFO][6763] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.155 [INFO][6763] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.155 [INFO][6763] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.159 [WARNING][6763] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.159 [INFO][6763] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" HandleID="k8s-pod-network.7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--apiserver--66c4f75c9f--qhkcw-eth0" Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.160 [INFO][6763] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.161647 containerd[1926]: 2025-11-08 01:39:46.160 [INFO][6746] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da" Nov 8 01:39:46.161950 containerd[1926]: time="2025-11-08T01:39:46.161665260Z" level=info msg="TearDown network for sandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\" successfully" Nov 8 01:39:46.173500 containerd[1926]: time="2025-11-08T01:39:46.173459960Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:46.173500 containerd[1926]: time="2025-11-08T01:39:46.173484347Z" level=info msg="RemovePodSandbox \"7d3325b9a496d42acdaf0bd5596aeb516aa4a23d8836bba92eb1ac3b5bdc84da\" returns successfully" Nov 8 01:39:46.173775 containerd[1926]: time="2025-11-08T01:39:46.173762189Z" level=info msg="StopPodSandbox for \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\"" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.191 [WARNING][6788] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.191 [INFO][6788] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.191 [INFO][6788] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" iface="eth0" netns="" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.191 [INFO][6788] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.191 [INFO][6788] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.200 [INFO][6806] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.200 [INFO][6806] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.200 [INFO][6806] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.204 [WARNING][6806] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.204 [INFO][6806] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.206 [INFO][6806] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.207422 containerd[1926]: 2025-11-08 01:39:46.206 [INFO][6788] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.207708 containerd[1926]: time="2025-11-08T01:39:46.207448894Z" level=info msg="TearDown network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\" successfully" Nov 8 01:39:46.207708 containerd[1926]: time="2025-11-08T01:39:46.207469476Z" level=info msg="StopPodSandbox for \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\" returns successfully" Nov 8 01:39:46.207749 containerd[1926]: time="2025-11-08T01:39:46.207740694Z" level=info msg="RemovePodSandbox for \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\"" Nov 8 01:39:46.207770 containerd[1926]: time="2025-11-08T01:39:46.207755536Z" level=info msg="Forcibly stopping sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\"" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.224 [WARNING][6830] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" WorkloadEndpoint="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.224 [INFO][6830] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.224 [INFO][6830] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" iface="eth0" netns="" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.224 [INFO][6830] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.224 [INFO][6830] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.236 [INFO][6847] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.236 [INFO][6847] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.236 [INFO][6847] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.242 [WARNING][6847] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.242 [INFO][6847] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" HandleID="k8s-pod-network.0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-whisker--5d6f95b8f7--6lzzn-eth0" Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.243 [INFO][6847] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.245836 containerd[1926]: 2025-11-08 01:39:46.244 [INFO][6830] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d" Nov 8 01:39:46.245836 containerd[1926]: time="2025-11-08T01:39:46.245801811Z" level=info msg="TearDown network for sandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\" successfully" Nov 8 01:39:46.247592 containerd[1926]: time="2025-11-08T01:39:46.247535254Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:46.247592 containerd[1926]: time="2025-11-08T01:39:46.247558958Z" level=info msg="RemovePodSandbox \"0b406db3c930403f651e580189b37da46e008cfa3a34e926ebed56c4895e6b5d\" returns successfully" Nov 8 01:39:46.247821 containerd[1926]: time="2025-11-08T01:39:46.247793773Z" level=info msg="StopPodSandbox for \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\"" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.265 [WARNING][6875] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0", GenerateName:"calico-kube-controllers-566b9cf58b-", Namespace:"calico-system", SelfLink:"", UID:"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01", ResourceVersion:"996", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"566b9cf58b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4", Pod:"calico-kube-controllers-566b9cf58b-whc7z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"caliec314f3d499", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.265 [INFO][6875] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.265 [INFO][6875] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" iface="eth0" netns="" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.266 [INFO][6875] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.266 [INFO][6875] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.276 [INFO][6891] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.276 [INFO][6891] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.276 [INFO][6891] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.279 [WARNING][6891] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.279 [INFO][6891] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.281 [INFO][6891] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.282378 containerd[1926]: 2025-11-08 01:39:46.281 [INFO][6875] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.282378 containerd[1926]: time="2025-11-08T01:39:46.282372646Z" level=info msg="TearDown network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\" successfully" Nov 8 01:39:46.282677 containerd[1926]: time="2025-11-08T01:39:46.282392240Z" level=info msg="StopPodSandbox for \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\" returns successfully" Nov 8 01:39:46.282699 containerd[1926]: time="2025-11-08T01:39:46.282680287Z" level=info msg="RemovePodSandbox for \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\"" Nov 8 01:39:46.282717 containerd[1926]: time="2025-11-08T01:39:46.282697423Z" level=info msg="Forcibly stopping sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\"" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.300 [WARNING][6916] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0", GenerateName:"calico-kube-controllers-566b9cf58b-", Namespace:"calico-system", SelfLink:"", UID:"f5773a4d-1a91-4ae3-82ea-bf0402f8aa01", ResourceVersion:"996", Generation:0, CreationTimestamp:time.Date(2025, time.November, 8, 1, 39, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"566b9cf58b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-3863d1ecf5", ContainerID:"6aca255c63c30f0b51b47bdb7eb5199f321dbc1e15abb03d7d3ae7ff9bd2a1b4", Pod:"calico-kube-controllers-566b9cf58b-whc7z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"caliec314f3d499", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.300 [INFO][6916] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.300 [INFO][6916] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" iface="eth0" netns="" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.300 [INFO][6916] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.300 [INFO][6916] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.311 [INFO][6930] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.311 [INFO][6930] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.311 [INFO][6930] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.315 [WARNING][6930] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.315 [INFO][6930] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" HandleID="k8s-pod-network.45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Workload="ci--4081.3.6--n--3863d1ecf5-k8s-calico--kube--controllers--566b9cf58b--whc7z-eth0" Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.317 [INFO][6930] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Nov 8 01:39:46.318723 containerd[1926]: 2025-11-08 01:39:46.317 [INFO][6916] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff" Nov 8 01:39:46.318723 containerd[1926]: time="2025-11-08T01:39:46.318718334Z" level=info msg="TearDown network for sandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\" successfully" Nov 8 01:39:46.320260 containerd[1926]: time="2025-11-08T01:39:46.320203902Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Nov 8 01:39:46.320260 containerd[1926]: time="2025-11-08T01:39:46.320227660Z" level=info msg="RemovePodSandbox \"45677a334745e3030c0717e2f63f4aa7a911442a482cc9391c5a9261eac0c7ff\" returns successfully" Nov 8 01:39:49.711724 kubelet[3256]: E1108 01:39:49.711682 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:39:52.712012 kubelet[3256]: E1108 01:39:52.711960 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:39:52.712012 kubelet[3256]: E1108 01:39:52.711975 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:39:54.711848 kubelet[3256]: E1108 01:39:54.711813 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:39:55.711916 kubelet[3256]: E1108 01:39:55.711887 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:39:59.714323 containerd[1926]: time="2025-11-08T01:39:59.714184572Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 8 01:40:00.103772 containerd[1926]: time="2025-11-08T01:40:00.103715322Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:00.104216 containerd[1926]: time="2025-11-08T01:40:00.104154302Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 8 01:40:00.104271 containerd[1926]: time="2025-11-08T01:40:00.104232209Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 8 01:40:00.104500 kubelet[3256]: E1108 01:40:00.104446 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:40:00.104500 kubelet[3256]: E1108 01:40:00.104482 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:40:00.104818 kubelet[3256]: E1108 01:40:00.104576 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:52eaa7a533b1456aa7b1871d4874688d,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:00.106174 containerd[1926]: time="2025-11-08T01:40:00.106154246Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 8 01:40:00.471932 containerd[1926]: time="2025-11-08T01:40:00.471829107Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:00.472296 containerd[1926]: time="2025-11-08T01:40:00.472274565Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 8 01:40:00.472343 containerd[1926]: time="2025-11-08T01:40:00.472319714Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 8 01:40:00.472486 kubelet[3256]: E1108 01:40:00.472434 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:40:00.472486 kubelet[3256]: E1108 01:40:00.472469 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:40:00.472608 kubelet[3256]: E1108 01:40:00.472537 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:00.473705 kubelet[3256]: E1108 01:40:00.473688 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:40:04.716550 containerd[1926]: time="2025-11-08T01:40:04.715606277Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:40:05.069082 containerd[1926]: time="2025-11-08T01:40:05.068950632Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:05.082282 containerd[1926]: time="2025-11-08T01:40:05.082236625Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:40:05.082353 containerd[1926]: time="2025-11-08T01:40:05.082309538Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:40:05.082474 kubelet[3256]: E1108 01:40:05.082405 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:05.082474 kubelet[3256]: E1108 01:40:05.082454 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:05.082794 kubelet[3256]: E1108 01:40:05.082667 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv77z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:05.082927 containerd[1926]: time="2025-11-08T01:40:05.082690394Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 8 01:40:05.083801 kubelet[3256]: E1108 01:40:05.083785 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:40:05.460258 containerd[1926]: time="2025-11-08T01:40:05.459978935Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:05.461005 containerd[1926]: time="2025-11-08T01:40:05.460926622Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 8 01:40:05.461036 containerd[1926]: time="2025-11-08T01:40:05.460995594Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 8 01:40:05.461126 kubelet[3256]: E1108 01:40:05.461102 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:40:05.461155 kubelet[3256]: E1108 01:40:05.461136 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:40:05.461262 kubelet[3256]: E1108 01:40:05.461209 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cts9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:05.462971 kubelet[3256]: E1108 01:40:05.462929 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:40:05.714705 containerd[1926]: time="2025-11-08T01:40:05.714527115Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:40:06.067368 containerd[1926]: time="2025-11-08T01:40:06.067308508Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:06.067829 containerd[1926]: time="2025-11-08T01:40:06.067767316Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:40:06.067876 containerd[1926]: time="2025-11-08T01:40:06.067831674Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:40:06.067982 kubelet[3256]: E1108 01:40:06.067932 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:06.067982 kubelet[3256]: E1108 01:40:06.067964 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:06.068074 kubelet[3256]: E1108 01:40:06.068044 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b49rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:06.069370 kubelet[3256]: E1108 01:40:06.069320 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:40:07.714132 containerd[1926]: time="2025-11-08T01:40:07.714028136Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 8 01:40:08.048250 containerd[1926]: time="2025-11-08T01:40:08.048163706Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:08.049081 containerd[1926]: time="2025-11-08T01:40:08.049054662Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 8 01:40:08.049137 containerd[1926]: time="2025-11-08T01:40:08.049116350Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 8 01:40:08.049230 kubelet[3256]: E1108 01:40:08.049208 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:40:08.049412 kubelet[3256]: E1108 01:40:08.049241 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:40:08.049412 kubelet[3256]: E1108 01:40:08.049322 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rplv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:08.050499 kubelet[3256]: E1108 01:40:08.050485 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:40:09.711731 containerd[1926]: time="2025-11-08T01:40:09.711674860Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 8 01:40:10.087922 containerd[1926]: time="2025-11-08T01:40:10.087895082Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:10.088415 containerd[1926]: time="2025-11-08T01:40:10.088368877Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 8 01:40:10.088488 containerd[1926]: time="2025-11-08T01:40:10.088414062Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 8 01:40:10.088586 kubelet[3256]: E1108 01:40:10.088536 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:40:10.088586 kubelet[3256]: E1108 01:40:10.088568 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:40:10.088851 kubelet[3256]: E1108 01:40:10.088641 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:10.090190 containerd[1926]: time="2025-11-08T01:40:10.090177076Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 8 01:40:10.454815 containerd[1926]: time="2025-11-08T01:40:10.454545412Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:10.455606 containerd[1926]: time="2025-11-08T01:40:10.455534208Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 8 01:40:10.455655 containerd[1926]: time="2025-11-08T01:40:10.455606934Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 8 01:40:10.455761 kubelet[3256]: E1108 01:40:10.455710 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:40:10.455761 kubelet[3256]: E1108 01:40:10.455740 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:40:10.455865 kubelet[3256]: E1108 01:40:10.455809 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:10.457046 kubelet[3256]: E1108 01:40:10.456994 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:40:13.713748 kubelet[3256]: E1108 01:40:13.713645 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:40:18.714009 kubelet[3256]: E1108 01:40:18.713890 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:40:20.712122 kubelet[3256]: E1108 01:40:20.712079 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:40:20.712544 kubelet[3256]: E1108 01:40:20.712360 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:40:21.712212 kubelet[3256]: E1108 01:40:21.712187 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:40:21.712212 kubelet[3256]: E1108 01:40:21.712187 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:40:27.712231 kubelet[3256]: E1108 01:40:27.712207 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:40:31.713921 kubelet[3256]: E1108 01:40:31.713775 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:40:31.715414 kubelet[3256]: E1108 01:40:31.714850 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:40:32.712200 kubelet[3256]: E1108 01:40:32.712157 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:40:33.713120 kubelet[3256]: E1108 01:40:33.713021 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:40:35.712579 kubelet[3256]: E1108 01:40:35.712558 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:40:40.714698 containerd[1926]: time="2025-11-08T01:40:40.714581497Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 8 01:40:41.082415 containerd[1926]: time="2025-11-08T01:40:41.082287182Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:41.083202 containerd[1926]: time="2025-11-08T01:40:41.083094315Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 8 01:40:41.083202 containerd[1926]: time="2025-11-08T01:40:41.083149467Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 8 01:40:41.083320 kubelet[3256]: E1108 01:40:41.083282 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:40:41.083533 kubelet[3256]: E1108 01:40:41.083325 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:40:41.083533 kubelet[3256]: E1108 01:40:41.083408 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:52eaa7a533b1456aa7b1871d4874688d,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:41.085001 containerd[1926]: time="2025-11-08T01:40:41.084988834Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 8 01:40:41.469220 containerd[1926]: time="2025-11-08T01:40:41.465247107Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:41.469220 containerd[1926]: time="2025-11-08T01:40:41.466362326Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 8 01:40:41.469220 containerd[1926]: time="2025-11-08T01:40:41.466445250Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 8 01:40:41.469433 kubelet[3256]: E1108 01:40:41.469263 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:40:41.469433 kubelet[3256]: E1108 01:40:41.469352 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:40:41.469523 kubelet[3256]: E1108 01:40:41.469463 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:41.470718 kubelet[3256]: E1108 01:40:41.470653 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:40:44.711874 kubelet[3256]: E1108 01:40:44.711808 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:40:44.711874 kubelet[3256]: E1108 01:40:44.711836 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:40:44.712413 kubelet[3256]: E1108 01:40:44.712061 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:40:46.712131 kubelet[3256]: E1108 01:40:46.712101 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:40:46.712458 containerd[1926]: time="2025-11-08T01:40:46.712211302Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 8 01:40:47.094106 containerd[1926]: time="2025-11-08T01:40:47.093968830Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:47.094933 containerd[1926]: time="2025-11-08T01:40:47.094904305Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 8 01:40:47.095009 containerd[1926]: time="2025-11-08T01:40:47.094974828Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 8 01:40:47.095046 kubelet[3256]: E1108 01:40:47.095024 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:40:47.095078 kubelet[3256]: E1108 01:40:47.095052 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:40:47.095203 kubelet[3256]: E1108 01:40:47.095132 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cts9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:47.096293 kubelet[3256]: E1108 01:40:47.096250 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:40:55.143005 update_engine[1914]: I20251108 01:40:55.142946 1914 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Nov 8 01:40:55.143005 update_engine[1914]: I20251108 01:40:55.142975 1914 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Nov 8 01:40:55.143308 update_engine[1914]: I20251108 01:40:55.143079 1914 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Nov 8 01:40:55.143342 update_engine[1914]: I20251108 01:40:55.143329 1914 omaha_request_params.cc:62] Current group set to lts Nov 8 01:40:55.143433 update_engine[1914]: I20251108 01:40:55.143390 1914 update_attempter.cc:499] Already updated boot flags. Skipping. Nov 8 01:40:55.143433 update_engine[1914]: I20251108 01:40:55.143399 1914 update_attempter.cc:643] Scheduling an action processor start. Nov 8 01:40:55.143433 update_engine[1914]: I20251108 01:40:55.143407 1914 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Nov 8 01:40:55.143433 update_engine[1914]: I20251108 01:40:55.143424 1914 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Nov 8 01:40:55.143544 update_engine[1914]: I20251108 01:40:55.143460 1914 omaha_request_action.cc:271] Posting an Omaha request to disabled Nov 8 01:40:55.143544 update_engine[1914]: I20251108 01:40:55.143466 1914 omaha_request_action.cc:272] Request: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: Nov 8 01:40:55.143544 update_engine[1914]: I20251108 01:40:55.143471 1914 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Nov 8 01:40:55.143738 locksmithd[1969]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Nov 8 01:40:55.144325 update_engine[1914]: I20251108 01:40:55.144285 1914 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Nov 8 01:40:55.144497 update_engine[1914]: I20251108 01:40:55.144455 1914 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Nov 8 01:40:55.145193 update_engine[1914]: E20251108 01:40:55.145169 1914 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Nov 8 01:40:55.145231 update_engine[1914]: I20251108 01:40:55.145207 1914 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Nov 8 01:40:55.712953 containerd[1926]: time="2025-11-08T01:40:55.712929042Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:40:56.071863 containerd[1926]: time="2025-11-08T01:40:56.071784095Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:56.072560 containerd[1926]: time="2025-11-08T01:40:56.072474328Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:40:56.072634 containerd[1926]: time="2025-11-08T01:40:56.072539395Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:40:56.072719 kubelet[3256]: E1108 01:40:56.072656 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:56.072719 kubelet[3256]: E1108 01:40:56.072698 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:56.072972 kubelet[3256]: E1108 01:40:56.072799 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b49rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:56.074066 kubelet[3256]: E1108 01:40:56.073998 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:40:56.714996 containerd[1926]: time="2025-11-08T01:40:56.714901573Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:40:56.715761 kubelet[3256]: E1108 01:40:56.715480 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:40:57.063798 containerd[1926]: time="2025-11-08T01:40:57.063698449Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:40:57.064794 containerd[1926]: time="2025-11-08T01:40:57.064767938Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:40:57.064868 containerd[1926]: time="2025-11-08T01:40:57.064841634Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:40:57.065011 kubelet[3256]: E1108 01:40:57.064952 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:57.065011 kubelet[3256]: E1108 01:40:57.065008 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:40:57.065137 kubelet[3256]: E1108 01:40:57.065114 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv77z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:40:57.066441 kubelet[3256]: E1108 01:40:57.066411 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:40:59.711703 containerd[1926]: time="2025-11-08T01:40:59.711682910Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 8 01:41:00.046802 containerd[1926]: time="2025-11-08T01:41:00.046679806Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:41:00.047677 containerd[1926]: time="2025-11-08T01:41:00.047594601Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 8 01:41:00.047677 containerd[1926]: time="2025-11-08T01:41:00.047660201Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 8 01:41:00.047786 kubelet[3256]: E1108 01:41:00.047762 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:41:00.047978 kubelet[3256]: E1108 01:41:00.047795 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:41:00.047978 kubelet[3256]: E1108 01:41:00.047862 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 8 01:41:00.049633 containerd[1926]: time="2025-11-08T01:41:00.049606064Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 8 01:41:00.399666 containerd[1926]: time="2025-11-08T01:41:00.399404271Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:41:00.400439 containerd[1926]: time="2025-11-08T01:41:00.400355869Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 8 01:41:00.400439 containerd[1926]: time="2025-11-08T01:41:00.400424352Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 8 01:41:00.400639 kubelet[3256]: E1108 01:41:00.400580 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:41:00.400639 kubelet[3256]: E1108 01:41:00.400620 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:41:00.400776 kubelet[3256]: E1108 01:41:00.400710 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 8 01:41:00.401928 kubelet[3256]: E1108 01:41:00.401888 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:41:01.712286 kubelet[3256]: E1108 01:41:01.712241 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:41:01.713150 containerd[1926]: time="2025-11-08T01:41:01.712472807Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 8 01:41:02.081455 containerd[1926]: time="2025-11-08T01:41:02.081361871Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:41:02.082209 containerd[1926]: time="2025-11-08T01:41:02.082187656Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 8 01:41:02.082278 containerd[1926]: time="2025-11-08T01:41:02.082270711Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 8 01:41:02.082510 kubelet[3256]: E1108 01:41:02.082450 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:41:02.082538 kubelet[3256]: E1108 01:41:02.082521 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:41:02.082655 kubelet[3256]: E1108 01:41:02.082612 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rplv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 8 01:41:02.084312 kubelet[3256]: E1108 01:41:02.084269 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:41:05.111447 update_engine[1914]: I20251108 01:41:05.111281 1914 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Nov 8 01:41:05.112406 update_engine[1914]: I20251108 01:41:05.111815 1914 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Nov 8 01:41:05.112406 update_engine[1914]: I20251108 01:41:05.112329 1914 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Nov 8 01:41:05.113361 update_engine[1914]: E20251108 01:41:05.113250 1914 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Nov 8 01:41:05.113572 update_engine[1914]: I20251108 01:41:05.113397 1914 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Nov 8 01:41:07.712246 kubelet[3256]: E1108 01:41:07.712223 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:41:10.714286 kubelet[3256]: E1108 01:41:10.714166 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:41:11.714970 kubelet[3256]: E1108 01:41:11.714868 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:41:13.712354 kubelet[3256]: E1108 01:41:13.712321 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:41:15.111643 update_engine[1914]: I20251108 01:41:15.111521 1914 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Nov 8 01:41:15.112406 update_engine[1914]: I20251108 01:41:15.111927 1914 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Nov 8 01:41:15.112406 update_engine[1914]: I20251108 01:41:15.112332 1914 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Nov 8 01:41:15.113180 update_engine[1914]: E20251108 01:41:15.113080 1914 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Nov 8 01:41:15.113316 update_engine[1914]: I20251108 01:41:15.113210 1914 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Nov 8 01:41:15.711453 kubelet[3256]: E1108 01:41:15.711432 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:41:15.711746 kubelet[3256]: E1108 01:41:15.711458 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:41:19.712730 kubelet[3256]: E1108 01:41:19.712639 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:41:22.714402 kubelet[3256]: E1108 01:41:22.714318 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:41:25.111360 update_engine[1914]: I20251108 01:41:25.111236 1914 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Nov 8 01:41:25.112129 update_engine[1914]: I20251108 01:41:25.111631 1914 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Nov 8 01:41:25.112129 update_engine[1914]: I20251108 01:41:25.111991 1914 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Nov 8 01:41:25.112789 update_engine[1914]: E20251108 01:41:25.112701 1914 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Nov 8 01:41:25.112945 update_engine[1914]: I20251108 01:41:25.112802 1914 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Nov 8 01:41:25.112945 update_engine[1914]: I20251108 01:41:25.112826 1914 omaha_request_action.cc:617] Omaha request response: Nov 8 01:41:25.113095 update_engine[1914]: E20251108 01:41:25.112946 1914 omaha_request_action.cc:636] Omaha request network transfer failed. Nov 8 01:41:25.113095 update_engine[1914]: I20251108 01:41:25.112986 1914 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Nov 8 01:41:25.113095 update_engine[1914]: I20251108 01:41:25.113001 1914 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Nov 8 01:41:25.113095 update_engine[1914]: I20251108 01:41:25.113011 1914 update_attempter.cc:306] Processing Done. Nov 8 01:41:25.113095 update_engine[1914]: E20251108 01:41:25.113034 1914 update_attempter.cc:619] Update failed. Nov 8 01:41:25.113095 update_engine[1914]: I20251108 01:41:25.113046 1914 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Nov 8 01:41:25.113095 update_engine[1914]: I20251108 01:41:25.113058 1914 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Nov 8 01:41:25.113095 update_engine[1914]: I20251108 01:41:25.113070 1914 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Nov 8 01:41:25.113800 update_engine[1914]: I20251108 01:41:25.113209 1914 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Nov 8 01:41:25.113800 update_engine[1914]: I20251108 01:41:25.113261 1914 omaha_request_action.cc:271] Posting an Omaha request to disabled Nov 8 01:41:25.113800 update_engine[1914]: I20251108 01:41:25.113276 1914 omaha_request_action.cc:272] Request: Nov 8 01:41:25.113800 update_engine[1914]: Nov 8 01:41:25.113800 update_engine[1914]: Nov 8 01:41:25.113800 update_engine[1914]: Nov 8 01:41:25.113800 update_engine[1914]: Nov 8 01:41:25.113800 update_engine[1914]: Nov 8 01:41:25.113800 update_engine[1914]: Nov 8 01:41:25.113800 update_engine[1914]: I20251108 01:41:25.113288 1914 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Nov 8 01:41:25.113800 update_engine[1914]: I20251108 01:41:25.113569 1914 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Nov 8 01:41:25.114465 locksmithd[1969]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.113875 1914 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Nov 8 01:41:25.115011 update_engine[1914]: E20251108 01:41:25.114480 1914 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.114578 1914 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.114598 1914 omaha_request_action.cc:617] Omaha request response: Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.114611 1914 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.114621 1914 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.114631 1914 update_attempter.cc:306] Processing Done. Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.114643 1914 update_attempter.cc:310] Error event sent. Nov 8 01:41:25.115011 update_engine[1914]: I20251108 01:41:25.114663 1914 update_check_scheduler.cc:74] Next update check in 47m7s Nov 8 01:41:25.115557 locksmithd[1969]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Nov 8 01:41:25.717736 kubelet[3256]: E1108 01:41:25.717607 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:41:26.715204 kubelet[3256]: E1108 01:41:26.715092 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:41:26.715967 kubelet[3256]: E1108 01:41:26.715895 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:41:30.714483 kubelet[3256]: E1108 01:41:30.714388 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:41:31.711482 kubelet[3256]: E1108 01:41:31.711458 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:41:35.715341 kubelet[3256]: E1108 01:41:35.715262 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:41:37.712500 kubelet[3256]: E1108 01:41:37.712475 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:41:39.712107 kubelet[3256]: E1108 01:41:39.712074 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:41:41.712409 kubelet[3256]: E1108 01:41:41.712343 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:41:43.713729 kubelet[3256]: E1108 01:41:43.713594 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:41:45.712558 kubelet[3256]: E1108 01:41:45.712524 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:41:48.713563 kubelet[3256]: E1108 01:41:48.713450 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:41:50.712246 kubelet[3256]: E1108 01:41:50.712203 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:41:52.713039 kubelet[3256]: E1108 01:41:52.712985 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:41:56.712596 kubelet[3256]: E1108 01:41:56.712506 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:41:57.713760 kubelet[3256]: E1108 01:41:57.713665 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:41:57.713760 kubelet[3256]: E1108 01:41:57.713667 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:42:01.711467 kubelet[3256]: E1108 01:42:01.711413 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:42:03.713594 kubelet[3256]: E1108 01:42:03.713568 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:42:04.713339 containerd[1926]: time="2025-11-08T01:42:04.713245005Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 8 01:42:05.096762 containerd[1926]: time="2025-11-08T01:42:05.096664962Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:05.097541 containerd[1926]: time="2025-11-08T01:42:05.097514557Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 8 01:42:05.097602 containerd[1926]: time="2025-11-08T01:42:05.097584978Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 8 01:42:05.097710 kubelet[3256]: E1108 01:42:05.097686 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:42:05.098042 kubelet[3256]: E1108 01:42:05.097720 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:42:05.098042 kubelet[3256]: E1108 01:42:05.097807 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:52eaa7a533b1456aa7b1871d4874688d,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:05.099379 containerd[1926]: time="2025-11-08T01:42:05.099335698Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 8 01:42:05.464492 containerd[1926]: time="2025-11-08T01:42:05.464227860Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:05.465107 containerd[1926]: time="2025-11-08T01:42:05.465034882Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 8 01:42:05.465107 containerd[1926]: time="2025-11-08T01:42:05.465067685Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 8 01:42:05.465198 kubelet[3256]: E1108 01:42:05.465161 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:42:05.465231 kubelet[3256]: E1108 01:42:05.465203 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:42:05.465330 kubelet[3256]: E1108 01:42:05.465267 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:05.466424 kubelet[3256]: E1108 01:42:05.466379 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:42:07.718684 kubelet[3256]: E1108 01:42:07.718584 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:42:09.712017 kubelet[3256]: E1108 01:42:09.711957 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:42:10.713985 containerd[1926]: time="2025-11-08T01:42:10.713906813Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 8 01:42:11.080926 containerd[1926]: time="2025-11-08T01:42:11.080869969Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:11.081466 containerd[1926]: time="2025-11-08T01:42:11.081401684Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 8 01:42:11.081509 containerd[1926]: time="2025-11-08T01:42:11.081465975Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 8 01:42:11.081591 kubelet[3256]: E1108 01:42:11.081537 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:42:11.081591 kubelet[3256]: E1108 01:42:11.081570 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:42:11.081810 kubelet[3256]: E1108 01:42:11.081647 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cts9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:11.082827 kubelet[3256]: E1108 01:42:11.082783 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:42:13.713443 kubelet[3256]: E1108 01:42:13.713339 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:42:15.715104 kubelet[3256]: E1108 01:42:15.715005 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:42:19.712465 kubelet[3256]: E1108 01:42:19.712438 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:42:22.712053 containerd[1926]: time="2025-11-08T01:42:22.712010964Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:42:23.080851 containerd[1926]: time="2025-11-08T01:42:23.080711123Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:23.081794 containerd[1926]: time="2025-11-08T01:42:23.081721880Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:42:23.081833 containerd[1926]: time="2025-11-08T01:42:23.081782683Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:42:23.081929 kubelet[3256]: E1108 01:42:23.081875 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:42:23.081929 kubelet[3256]: E1108 01:42:23.081904 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:42:23.082152 kubelet[3256]: E1108 01:42:23.082042 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv77z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:23.082220 containerd[1926]: time="2025-11-08T01:42:23.082103950Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 8 01:42:23.083182 kubelet[3256]: E1108 01:42:23.083167 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:42:23.462335 containerd[1926]: time="2025-11-08T01:42:23.462044970Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:23.463108 containerd[1926]: time="2025-11-08T01:42:23.463082113Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 8 01:42:23.463183 containerd[1926]: time="2025-11-08T01:42:23.463151274Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 8 01:42:23.463293 kubelet[3256]: E1108 01:42:23.463272 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:42:23.463335 kubelet[3256]: E1108 01:42:23.463303 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:42:23.463405 kubelet[3256]: E1108 01:42:23.463380 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rplv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:23.464555 kubelet[3256]: E1108 01:42:23.464540 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:42:24.714917 kubelet[3256]: E1108 01:42:24.714788 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:42:26.715643 containerd[1926]: time="2025-11-08T01:42:26.715509218Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 8 01:42:27.100957 containerd[1926]: time="2025-11-08T01:42:27.100843489Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:27.101811 containerd[1926]: time="2025-11-08T01:42:27.101722263Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 8 01:42:27.101811 containerd[1926]: time="2025-11-08T01:42:27.101790582Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 8 01:42:27.101973 kubelet[3256]: E1108 01:42:27.101922 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:42:27.101973 kubelet[3256]: E1108 01:42:27.101955 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:42:27.102360 kubelet[3256]: E1108 01:42:27.102140 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:27.102462 containerd[1926]: time="2025-11-08T01:42:27.102235988Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:42:27.500782 containerd[1926]: time="2025-11-08T01:42:27.500508612Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:27.501490 containerd[1926]: time="2025-11-08T01:42:27.501415566Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:42:27.501576 containerd[1926]: time="2025-11-08T01:42:27.501481539Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:42:27.501611 kubelet[3256]: E1108 01:42:27.501590 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:42:27.501641 kubelet[3256]: E1108 01:42:27.501616 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:42:27.501791 kubelet[3256]: E1108 01:42:27.501767 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b49rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:27.501941 containerd[1926]: time="2025-11-08T01:42:27.501868513Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 8 01:42:27.502992 kubelet[3256]: E1108 01:42:27.502976 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:42:27.869827 containerd[1926]: time="2025-11-08T01:42:27.869723682Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:42:27.870835 containerd[1926]: time="2025-11-08T01:42:27.870768525Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 8 01:42:27.870877 containerd[1926]: time="2025-11-08T01:42:27.870835413Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 8 01:42:27.870997 kubelet[3256]: E1108 01:42:27.870948 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:42:27.870997 kubelet[3256]: E1108 01:42:27.870981 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:42:27.871079 kubelet[3256]: E1108 01:42:27.871052 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 8 01:42:27.872376 kubelet[3256]: E1108 01:42:27.872331 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:42:32.712117 kubelet[3256]: E1108 01:42:32.712081 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:42:33.711643 kubelet[3256]: E1108 01:42:33.711614 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:42:36.713544 kubelet[3256]: E1108 01:42:36.713425 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:42:37.713650 kubelet[3256]: E1108 01:42:37.713568 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:42:38.712008 kubelet[3256]: E1108 01:42:38.711951 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:42:40.713685 kubelet[3256]: E1108 01:42:40.713626 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:42:44.712988 kubelet[3256]: E1108 01:42:44.712890 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:42:45.712223 kubelet[3256]: E1108 01:42:45.712198 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:42:50.713480 kubelet[3256]: E1108 01:42:50.713367 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:42:51.712164 kubelet[3256]: E1108 01:42:51.712131 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:42:52.712292 kubelet[3256]: E1108 01:42:52.712212 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:42:55.715868 kubelet[3256]: E1108 01:42:55.715701 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:42:56.713756 kubelet[3256]: E1108 01:42:56.713661 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:42:56.714726 kubelet[3256]: E1108 01:42:56.714584 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:43:01.716788 kubelet[3256]: E1108 01:43:01.716729 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:43:04.711904 kubelet[3256]: E1108 01:43:04.711848 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:43:05.713692 kubelet[3256]: E1108 01:43:05.713666 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:43:07.712386 kubelet[3256]: E1108 01:43:07.712361 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:43:09.712805 kubelet[3256]: E1108 01:43:09.712735 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:43:10.712407 kubelet[3256]: E1108 01:43:10.712350 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:43:15.711718 kubelet[3256]: E1108 01:43:15.711696 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:43:19.713403 kubelet[3256]: E1108 01:43:19.713302 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:43:20.713985 kubelet[3256]: E1108 01:43:20.713905 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:43:20.714956 kubelet[3256]: E1108 01:43:20.714628 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:43:21.711663 kubelet[3256]: E1108 01:43:21.711640 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:43:23.712080 kubelet[3256]: E1108 01:43:23.712054 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:43:27.713901 kubelet[3256]: E1108 01:43:27.713794 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:43:32.712490 kubelet[3256]: E1108 01:43:32.712456 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:43:32.712956 kubelet[3256]: E1108 01:43:32.712769 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:43:33.713354 kubelet[3256]: E1108 01:43:33.713303 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:43:34.713566 kubelet[3256]: E1108 01:43:34.713462 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:43:37.712535 kubelet[3256]: E1108 01:43:37.712498 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:43:41.713181 kubelet[3256]: E1108 01:43:41.713064 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:43:43.714477 kubelet[3256]: E1108 01:43:43.714390 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:43:45.712930 kubelet[3256]: E1108 01:43:45.712877 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:43:46.713489 kubelet[3256]: E1108 01:43:46.713395 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:43:46.714752 kubelet[3256]: E1108 01:43:46.713694 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:43:49.714561 kubelet[3256]: E1108 01:43:49.714478 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:43:52.713714 kubelet[3256]: E1108 01:43:52.713613 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:43:55.712473 kubelet[3256]: E1108 01:43:55.712449 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:43:57.714785 kubelet[3256]: E1108 01:43:57.714667 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:44:00.711723 kubelet[3256]: E1108 01:44:00.711679 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:44:00.712150 kubelet[3256]: E1108 01:44:00.711738 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:44:04.713553 kubelet[3256]: E1108 01:44:04.713491 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:44:05.714066 kubelet[3256]: E1108 01:44:05.713985 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:44:09.711750 kubelet[3256]: E1108 01:44:09.711691 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:44:11.712271 kubelet[3256]: E1108 01:44:11.712213 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:44:12.712043 kubelet[3256]: E1108 01:44:12.711996 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:44:14.712432 kubelet[3256]: E1108 01:44:14.712376 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:44:18.714902 kubelet[3256]: E1108 01:44:18.714774 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:44:20.712621 kubelet[3256]: E1108 01:44:20.712518 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:44:22.713372 kubelet[3256]: E1108 01:44:22.713249 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:44:24.712516 kubelet[3256]: E1108 01:44:24.712485 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:44:26.711724 kubelet[3256]: E1108 01:44:26.711660 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:44:27.714079 kubelet[3256]: E1108 01:44:27.713959 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:44:30.712127 kubelet[3256]: E1108 01:44:30.712089 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:44:34.715047 kubelet[3256]: E1108 01:44:34.714935 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:44:34.715047 kubelet[3256]: E1108 01:44:34.714945 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:44:36.712570 kubelet[3256]: E1108 01:44:36.712511 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:44:39.715319 kubelet[3256]: E1108 01:44:39.715191 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:44:41.277453 systemd[1]: Started sshd@9-139.178.94.189:22-139.178.68.195:43978.service - OpenSSH per-connection server daemon (139.178.68.195:43978). Nov 8 01:44:41.306396 sshd[7417]: Accepted publickey for core from 139.178.68.195 port 43978 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:44:41.307246 sshd[7417]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:44:41.310165 systemd-logind[1909]: New session 12 of user core. Nov 8 01:44:41.326408 systemd[1]: Started session-12.scope - Session 12 of User core. Nov 8 01:44:41.452186 sshd[7417]: pam_unix(sshd:session): session closed for user core Nov 8 01:44:41.455617 systemd[1]: sshd@9-139.178.94.189:22-139.178.68.195:43978.service: Deactivated successfully. Nov 8 01:44:41.457740 systemd-logind[1909]: Session 12 logged out. Waiting for processes to exit. Nov 8 01:44:41.457801 systemd[1]: session-12.scope: Deactivated successfully. Nov 8 01:44:41.458961 systemd-logind[1909]: Removed session 12. Nov 8 01:44:42.717373 kubelet[3256]: E1108 01:44:42.717347 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:44:42.717773 kubelet[3256]: E1108 01:44:42.717563 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:44:46.478485 systemd[1]: Started sshd@10-139.178.94.189:22-139.178.68.195:43676.service - OpenSSH per-connection server daemon (139.178.68.195:43676). Nov 8 01:44:46.518305 sshd[7460]: Accepted publickey for core from 139.178.68.195 port 43676 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:44:46.519272 sshd[7460]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:44:46.522402 systemd-logind[1909]: New session 13 of user core. Nov 8 01:44:46.534357 systemd[1]: Started session-13.scope - Session 13 of User core. Nov 8 01:44:46.614303 sshd[7460]: pam_unix(sshd:session): session closed for user core Nov 8 01:44:46.615938 systemd[1]: sshd@10-139.178.94.189:22-139.178.68.195:43676.service: Deactivated successfully. Nov 8 01:44:46.617378 systemd-logind[1909]: Session 13 logged out. Waiting for processes to exit. Nov 8 01:44:46.617418 systemd[1]: session-13.scope: Deactivated successfully. Nov 8 01:44:46.618028 systemd-logind[1909]: Removed session 13. Nov 8 01:44:46.713625 kubelet[3256]: E1108 01:44:46.713517 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:44:48.711455 kubelet[3256]: E1108 01:44:48.711407 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:44:48.711455 kubelet[3256]: E1108 01:44:48.711407 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:44:51.628358 systemd[1]: Started sshd@11-139.178.94.189:22-139.178.68.195:43680.service - OpenSSH per-connection server daemon (139.178.68.195:43680). Nov 8 01:44:51.655608 sshd[7524]: Accepted publickey for core from 139.178.68.195 port 43680 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:44:51.656411 sshd[7524]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:44:51.658942 systemd-logind[1909]: New session 14 of user core. Nov 8 01:44:51.671448 systemd[1]: Started session-14.scope - Session 14 of User core. Nov 8 01:44:51.752427 sshd[7524]: pam_unix(sshd:session): session closed for user core Nov 8 01:44:51.769385 systemd[1]: Started sshd@12-139.178.94.189:22-139.178.68.195:43692.service - OpenSSH per-connection server daemon (139.178.68.195:43692). Nov 8 01:44:51.769741 systemd[1]: sshd@11-139.178.94.189:22-139.178.68.195:43680.service: Deactivated successfully. Nov 8 01:44:51.770665 systemd[1]: session-14.scope: Deactivated successfully. Nov 8 01:44:51.771359 systemd-logind[1909]: Session 14 logged out. Waiting for processes to exit. Nov 8 01:44:51.771948 systemd-logind[1909]: Removed session 14. Nov 8 01:44:51.798708 sshd[7550]: Accepted publickey for core from 139.178.68.195 port 43692 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:44:51.799686 sshd[7550]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:44:51.802394 systemd-logind[1909]: New session 15 of user core. Nov 8 01:44:51.820450 systemd[1]: Started session-15.scope - Session 15 of User core. Nov 8 01:44:51.973478 sshd[7550]: pam_unix(sshd:session): session closed for user core Nov 8 01:44:51.982357 systemd[1]: Started sshd@13-139.178.94.189:22-139.178.68.195:43696.service - OpenSSH per-connection server daemon (139.178.68.195:43696). Nov 8 01:44:51.982749 systemd[1]: sshd@12-139.178.94.189:22-139.178.68.195:43692.service: Deactivated successfully. Nov 8 01:44:51.984053 systemd[1]: session-15.scope: Deactivated successfully. Nov 8 01:44:51.984958 systemd-logind[1909]: Session 15 logged out. Waiting for processes to exit. Nov 8 01:44:51.985539 systemd-logind[1909]: Removed session 15. Nov 8 01:44:52.010370 sshd[7576]: Accepted publickey for core from 139.178.68.195 port 43696 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:44:52.011157 sshd[7576]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:44:52.013702 systemd-logind[1909]: New session 16 of user core. Nov 8 01:44:52.031328 systemd[1]: Started session-16.scope - Session 16 of User core. Nov 8 01:44:52.164912 sshd[7576]: pam_unix(sshd:session): session closed for user core Nov 8 01:44:52.167103 systemd[1]: sshd@13-139.178.94.189:22-139.178.68.195:43696.service: Deactivated successfully. Nov 8 01:44:52.169171 systemd-logind[1909]: Session 16 logged out. Waiting for processes to exit. Nov 8 01:44:52.169328 systemd[1]: session-16.scope: Deactivated successfully. Nov 8 01:44:52.170118 systemd-logind[1909]: Removed session 16. Nov 8 01:44:52.713712 kubelet[3256]: E1108 01:44:52.713630 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:44:57.189810 systemd[1]: Started sshd@14-139.178.94.189:22-139.178.68.195:42998.service - OpenSSH per-connection server daemon (139.178.68.195:42998). Nov 8 01:44:57.248055 sshd[7606]: Accepted publickey for core from 139.178.68.195 port 42998 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:44:57.248987 sshd[7606]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:44:57.251698 systemd-logind[1909]: New session 17 of user core. Nov 8 01:44:57.268435 systemd[1]: Started session-17.scope - Session 17 of User core. Nov 8 01:44:57.376579 sshd[7606]: pam_unix(sshd:session): session closed for user core Nov 8 01:44:57.378220 systemd[1]: sshd@14-139.178.94.189:22-139.178.68.195:42998.service: Deactivated successfully. Nov 8 01:44:57.379842 systemd-logind[1909]: Session 17 logged out. Waiting for processes to exit. Nov 8 01:44:57.379935 systemd[1]: session-17.scope: Deactivated successfully. Nov 8 01:44:57.380609 systemd-logind[1909]: Removed session 17. Nov 8 01:44:57.711910 containerd[1926]: time="2025-11-08T01:44:57.711849712Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Nov 8 01:44:58.079018 containerd[1926]: time="2025-11-08T01:44:58.078991359Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:44:58.079467 containerd[1926]: time="2025-11-08T01:44:58.079445705Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Nov 8 01:44:58.079497 containerd[1926]: time="2025-11-08T01:44:58.079469812Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Nov 8 01:44:58.079631 kubelet[3256]: E1108 01:44:58.079574 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:44:58.079922 kubelet[3256]: E1108 01:44:58.079659 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Nov 8 01:44:58.079922 kubelet[3256]: E1108 01:44:58.079839 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cts9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-566b9cf58b-whc7z_calico-system(f5773a4d-1a91-4ae3-82ea-bf0402f8aa01): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Nov 8 01:44:58.080042 containerd[1926]: time="2025-11-08T01:44:58.079860029Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Nov 8 01:44:58.081011 kubelet[3256]: E1108 01:44:58.080996 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:44:58.439428 containerd[1926]: time="2025-11-08T01:44:58.439136337Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:44:58.440204 containerd[1926]: time="2025-11-08T01:44:58.440100650Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Nov 8 01:44:58.440260 containerd[1926]: time="2025-11-08T01:44:58.440175220Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Nov 8 01:44:58.440423 kubelet[3256]: E1108 01:44:58.440372 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:44:58.440423 kubelet[3256]: E1108 01:44:58.440404 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Nov 8 01:44:58.440487 kubelet[3256]: E1108 01:44:58.440466 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:52eaa7a533b1456aa7b1871d4874688d,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Nov 8 01:44:58.442052 containerd[1926]: time="2025-11-08T01:44:58.442024675Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Nov 8 01:44:58.783193 containerd[1926]: time="2025-11-08T01:44:58.783089454Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:44:58.783995 containerd[1926]: time="2025-11-08T01:44:58.783896016Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Nov 8 01:44:58.783995 containerd[1926]: time="2025-11-08T01:44:58.783955473Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Nov 8 01:44:58.784072 kubelet[3256]: E1108 01:44:58.784030 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:44:58.784072 kubelet[3256]: E1108 01:44:58.784058 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Nov 8 01:44:58.784191 kubelet[3256]: E1108 01:44:58.784128 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqv8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c9bb6b6fc-xlbx8_calico-system(8ffe5d99-6a33-4ea1-b47d-a065e9dc4970): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Nov 8 01:44:58.785318 kubelet[3256]: E1108 01:44:58.785276 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:44:59.712193 kubelet[3256]: E1108 01:44:59.712172 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:45:00.711608 kubelet[3256]: E1108 01:45:00.711566 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:45:01.713501 kubelet[3256]: E1108 01:45:01.713408 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:45:02.388857 systemd[1]: Started sshd@15-139.178.94.189:22-139.178.68.195:43008.service - OpenSSH per-connection server daemon (139.178.68.195:43008). Nov 8 01:45:02.420623 sshd[7637]: Accepted publickey for core from 139.178.68.195 port 43008 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:02.421318 sshd[7637]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:02.423891 systemd-logind[1909]: New session 18 of user core. Nov 8 01:45:02.432354 systemd[1]: Started session-18.scope - Session 18 of User core. Nov 8 01:45:02.515972 sshd[7637]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:02.517472 systemd[1]: sshd@15-139.178.94.189:22-139.178.68.195:43008.service: Deactivated successfully. Nov 8 01:45:02.518950 systemd-logind[1909]: Session 18 logged out. Waiting for processes to exit. Nov 8 01:45:02.518974 systemd[1]: session-18.scope: Deactivated successfully. Nov 8 01:45:02.519556 systemd-logind[1909]: Removed session 18. Nov 8 01:45:03.714887 kubelet[3256]: E1108 01:45:03.714731 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:45:07.531349 systemd[1]: Started sshd@16-139.178.94.189:22-139.178.68.195:47030.service - OpenSSH per-connection server daemon (139.178.68.195:47030). Nov 8 01:45:07.557517 sshd[7664]: Accepted publickey for core from 139.178.68.195 port 47030 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:07.558309 sshd[7664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:07.560834 systemd-logind[1909]: New session 19 of user core. Nov 8 01:45:07.561492 systemd[1]: Started session-19.scope - Session 19 of User core. Nov 8 01:45:07.646269 sshd[7664]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:07.648218 systemd[1]: sshd@16-139.178.94.189:22-139.178.68.195:47030.service: Deactivated successfully. Nov 8 01:45:07.649204 systemd-logind[1909]: Session 19 logged out. Waiting for processes to exit. Nov 8 01:45:07.649235 systemd[1]: session-19.scope: Deactivated successfully. Nov 8 01:45:07.649738 systemd-logind[1909]: Removed session 19. Nov 8 01:45:10.712451 containerd[1926]: time="2025-11-08T01:45:10.712367555Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Nov 8 01:45:11.094185 containerd[1926]: time="2025-11-08T01:45:11.094044693Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:45:11.095137 containerd[1926]: time="2025-11-08T01:45:11.095058663Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Nov 8 01:45:11.095189 containerd[1926]: time="2025-11-08T01:45:11.095129846Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Nov 8 01:45:11.095293 kubelet[3256]: E1108 01:45:11.095211 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:45:11.095293 kubelet[3256]: E1108 01:45:11.095252 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Nov 8 01:45:11.095499 kubelet[3256]: E1108 01:45:11.095354 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rplv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-8ld4r_calico-system(5f897823-41ac-4b66-8214-3b03dd8d6290): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Nov 8 01:45:11.096559 kubelet[3256]: E1108 01:45:11.096516 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:45:12.669544 systemd[1]: Started sshd@17-139.178.94.189:22-139.178.68.195:47036.service - OpenSSH per-connection server daemon (139.178.68.195:47036). Nov 8 01:45:12.711191 kubelet[3256]: E1108 01:45:12.711167 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:45:12.711468 containerd[1926]: time="2025-11-08T01:45:12.711293322Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:45:12.712592 sshd[7693]: Accepted publickey for core from 139.178.68.195 port 47036 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:12.713387 sshd[7693]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:12.715913 systemd-logind[1909]: New session 20 of user core. Nov 8 01:45:12.726369 systemd[1]: Started session-20.scope - Session 20 of User core. Nov 8 01:45:12.853564 sshd[7693]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:12.878454 systemd[1]: Started sshd@18-139.178.94.189:22-139.178.68.195:47038.service - OpenSSH per-connection server daemon (139.178.68.195:47038). Nov 8 01:45:12.878722 systemd[1]: sshd@17-139.178.94.189:22-139.178.68.195:47036.service: Deactivated successfully. Nov 8 01:45:12.879598 systemd[1]: session-20.scope: Deactivated successfully. Nov 8 01:45:12.880230 systemd-logind[1909]: Session 20 logged out. Waiting for processes to exit. Nov 8 01:45:12.880836 systemd-logind[1909]: Removed session 20. Nov 8 01:45:12.905601 sshd[7718]: Accepted publickey for core from 139.178.68.195 port 47038 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:12.906396 sshd[7718]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:12.909029 systemd-logind[1909]: New session 21 of user core. Nov 8 01:45:12.916334 systemd[1]: Started session-21.scope - Session 21 of User core. Nov 8 01:45:13.044846 containerd[1926]: time="2025-11-08T01:45:13.044815300Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:45:13.045320 containerd[1926]: time="2025-11-08T01:45:13.045295141Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:45:13.045378 containerd[1926]: time="2025-11-08T01:45:13.045364727Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:45:13.045487 kubelet[3256]: E1108 01:45:13.045461 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:45:13.045517 kubelet[3256]: E1108 01:45:13.045496 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:45:13.045596 kubelet[3256]: E1108 01:45:13.045572 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b49rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-gztt5_calico-apiserver(31ac439b-7263-4e83-932f-3464165d69ab): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:45:13.046865 kubelet[3256]: E1108 01:45:13.046849 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:45:13.084666 sshd[7718]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:13.101312 systemd[1]: Started sshd@19-139.178.94.189:22-139.178.68.195:35928.service - OpenSSH per-connection server daemon (139.178.68.195:35928). Nov 8 01:45:13.101666 systemd[1]: sshd@18-139.178.94.189:22-139.178.68.195:47038.service: Deactivated successfully. Nov 8 01:45:13.102854 systemd[1]: session-21.scope: Deactivated successfully. Nov 8 01:45:13.103780 systemd-logind[1909]: Session 21 logged out. Waiting for processes to exit. Nov 8 01:45:13.104527 systemd-logind[1909]: Removed session 21. Nov 8 01:45:13.148659 sshd[7744]: Accepted publickey for core from 139.178.68.195 port 35928 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:13.149864 sshd[7744]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:13.154063 systemd-logind[1909]: New session 22 of user core. Nov 8 01:45:13.176493 systemd[1]: Started session-22.scope - Session 22 of User core. Nov 8 01:45:13.712463 kubelet[3256]: E1108 01:45:13.712420 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:45:13.823043 sshd[7744]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:13.838403 systemd[1]: Started sshd@20-139.178.94.189:22-139.178.68.195:35934.service - OpenSSH per-connection server daemon (139.178.68.195:35934). Nov 8 01:45:13.838691 systemd[1]: sshd@19-139.178.94.189:22-139.178.68.195:35928.service: Deactivated successfully. Nov 8 01:45:13.839621 systemd[1]: session-22.scope: Deactivated successfully. Nov 8 01:45:13.840321 systemd-logind[1909]: Session 22 logged out. Waiting for processes to exit. Nov 8 01:45:13.841012 systemd-logind[1909]: Removed session 22. Nov 8 01:45:13.865314 sshd[7775]: Accepted publickey for core from 139.178.68.195 port 35934 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:13.866003 sshd[7775]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:13.868746 systemd-logind[1909]: New session 23 of user core. Nov 8 01:45:13.883291 systemd[1]: Started session-23.scope - Session 23 of User core. Nov 8 01:45:14.023055 sshd[7775]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:14.044404 systemd[1]: Started sshd@21-139.178.94.189:22-139.178.68.195:35942.service - OpenSSH per-connection server daemon (139.178.68.195:35942). Nov 8 01:45:14.044698 systemd[1]: sshd@20-139.178.94.189:22-139.178.68.195:35934.service: Deactivated successfully. Nov 8 01:45:14.045667 systemd[1]: session-23.scope: Deactivated successfully. Nov 8 01:45:14.046405 systemd-logind[1909]: Session 23 logged out. Waiting for processes to exit. Nov 8 01:45:14.047123 systemd-logind[1909]: Removed session 23. Nov 8 01:45:14.071502 sshd[7802]: Accepted publickey for core from 139.178.68.195 port 35942 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:14.072296 sshd[7802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:14.075017 systemd-logind[1909]: New session 24 of user core. Nov 8 01:45:14.090478 systemd[1]: Started session-24.scope - Session 24 of User core. Nov 8 01:45:14.167085 sshd[7802]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:14.168545 systemd[1]: sshd@21-139.178.94.189:22-139.178.68.195:35942.service: Deactivated successfully. Nov 8 01:45:14.169934 systemd-logind[1909]: Session 24 logged out. Waiting for processes to exit. Nov 8 01:45:14.169998 systemd[1]: session-24.scope: Deactivated successfully. Nov 8 01:45:14.170613 systemd-logind[1909]: Removed session 24. Nov 8 01:45:16.713218 containerd[1926]: time="2025-11-08T01:45:16.713121001Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Nov 8 01:45:17.065628 containerd[1926]: time="2025-11-08T01:45:17.065594883Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:45:17.069185 containerd[1926]: time="2025-11-08T01:45:17.069081694Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Nov 8 01:45:17.069239 containerd[1926]: time="2025-11-08T01:45:17.069182868Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Nov 8 01:45:17.069325 kubelet[3256]: E1108 01:45:17.069279 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:45:17.069325 kubelet[3256]: E1108 01:45:17.069314 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Nov 8 01:45:17.069673 kubelet[3256]: E1108 01:45:17.069386 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv77z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-66c4f75c9f-qhkcw_calico-apiserver(79b0696d-13d8-44db-9d91-6f1131e9b5ef): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Nov 8 01:45:17.070552 kubelet[3256]: E1108 01:45:17.070534 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:45:17.714426 containerd[1926]: time="2025-11-08T01:45:17.714333690Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Nov 8 01:45:18.072089 containerd[1926]: time="2025-11-08T01:45:18.071959391Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:45:18.073051 containerd[1926]: time="2025-11-08T01:45:18.072963615Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Nov 8 01:45:18.073051 containerd[1926]: time="2025-11-08T01:45:18.073026845Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Nov 8 01:45:18.073241 kubelet[3256]: E1108 01:45:18.073155 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:45:18.073241 kubelet[3256]: E1108 01:45:18.073217 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Nov 8 01:45:18.073498 kubelet[3256]: E1108 01:45:18.073285 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Nov 8 01:45:18.075004 containerd[1926]: time="2025-11-08T01:45:18.074992457Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Nov 8 01:45:18.469924 containerd[1926]: time="2025-11-08T01:45:18.469829916Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Nov 8 01:45:18.470249 containerd[1926]: time="2025-11-08T01:45:18.470156323Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Nov 8 01:45:18.470299 containerd[1926]: time="2025-11-08T01:45:18.470268477Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Nov 8 01:45:18.470416 kubelet[3256]: E1108 01:45:18.470370 3256 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:45:18.470416 kubelet[3256]: E1108 01:45:18.470402 3256 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Nov 8 01:45:18.470491 kubelet[3256]: E1108 01:45:18.470467 3256 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lrzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-dmqbg_calico-system(ff4b01f8-4bf6-4050-afe4-e2ac586eece3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Nov 8 01:45:18.471680 kubelet[3256]: E1108 01:45:18.471663 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-dmqbg" podUID="ff4b01f8-4bf6-4050-afe4-e2ac586eece3" Nov 8 01:45:19.187471 systemd[1]: Started sshd@22-139.178.94.189:22-139.178.68.195:35956.service - OpenSSH per-connection server daemon (139.178.68.195:35956). Nov 8 01:45:19.260808 sshd[7835]: Accepted publickey for core from 139.178.68.195 port 35956 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:19.262610 sshd[7835]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:19.268548 systemd-logind[1909]: New session 25 of user core. Nov 8 01:45:19.276584 systemd[1]: Started session-25.scope - Session 25 of User core. Nov 8 01:45:19.409326 sshd[7835]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:19.411237 systemd[1]: sshd@22-139.178.94.189:22-139.178.68.195:35956.service: Deactivated successfully. Nov 8 01:45:19.413007 systemd-logind[1909]: Session 25 logged out. Waiting for processes to exit. Nov 8 01:45:19.413124 systemd[1]: session-25.scope: Deactivated successfully. Nov 8 01:45:19.413857 systemd-logind[1909]: Removed session 25. Nov 8 01:45:23.712235 kubelet[3256]: E1108 01:45:23.712184 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-566b9cf58b-whc7z" podUID="f5773a4d-1a91-4ae3-82ea-bf0402f8aa01" Nov 8 01:45:23.712235 kubelet[3256]: E1108 01:45:23.712184 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-8ld4r" podUID="5f897823-41ac-4b66-8214-3b03dd8d6290" Nov 8 01:45:24.419500 systemd[1]: Started sshd@23-139.178.94.189:22-139.178.68.195:54794.service - OpenSSH per-connection server daemon (139.178.68.195:54794). Nov 8 01:45:24.472559 sshd[7898]: Accepted publickey for core from 139.178.68.195 port 54794 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:24.473680 sshd[7898]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:24.477222 systemd-logind[1909]: New session 26 of user core. Nov 8 01:45:24.493945 systemd[1]: Started session-26.scope - Session 26 of User core. Nov 8 01:45:24.640261 sshd[7898]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:24.642096 systemd[1]: sshd@23-139.178.94.189:22-139.178.68.195:54794.service: Deactivated successfully. Nov 8 01:45:24.643821 systemd-logind[1909]: Session 26 logged out. Waiting for processes to exit. Nov 8 01:45:24.643945 systemd[1]: session-26.scope: Deactivated successfully. Nov 8 01:45:24.644722 systemd-logind[1909]: Removed session 26. Nov 8 01:45:25.714169 kubelet[3256]: E1108 01:45:25.714070 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-gztt5" podUID="31ac439b-7263-4e83-932f-3464165d69ab" Nov 8 01:45:27.715539 kubelet[3256]: E1108 01:45:27.715440 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c9bb6b6fc-xlbx8" podUID="8ffe5d99-6a33-4ea1-b47d-a065e9dc4970" Nov 8 01:45:28.714855 kubelet[3256]: E1108 01:45:28.714730 3256 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-66c4f75c9f-qhkcw" podUID="79b0696d-13d8-44db-9d91-6f1131e9b5ef" Nov 8 01:45:29.660345 systemd[1]: Started sshd@24-139.178.94.189:22-139.178.68.195:54802.service - OpenSSH per-connection server daemon (139.178.68.195:54802). Nov 8 01:45:29.687105 sshd[7925]: Accepted publickey for core from 139.178.68.195 port 54802 ssh2: RSA SHA256:CDEH3Gh6VSwb5luG5uhujouIqwp740QGMGXihV+mnVQ Nov 8 01:45:29.687874 sshd[7925]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 8 01:45:29.690626 systemd-logind[1909]: New session 27 of user core. Nov 8 01:45:29.708398 systemd[1]: Started session-27.scope - Session 27 of User core. Nov 8 01:45:29.796306 sshd[7925]: pam_unix(sshd:session): session closed for user core Nov 8 01:45:29.797848 systemd[1]: sshd@24-139.178.94.189:22-139.178.68.195:54802.service: Deactivated successfully. Nov 8 01:45:29.799263 systemd-logind[1909]: Session 27 logged out. Waiting for processes to exit. Nov 8 01:45:29.799376 systemd[1]: session-27.scope: Deactivated successfully. Nov 8 01:45:29.799937 systemd-logind[1909]: Removed session 27.