Oct 13 00:01:33.830683 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Oct 13 00:01:33.830705 kernel: Linux version 6.12.51-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Sun Oct 12 22:32:01 -00 2025 Oct 13 00:01:33.830715 kernel: KASLR enabled Oct 13 00:01:33.830720 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Oct 13 00:01:33.830726 kernel: efi: SMBIOS 3.0=0x139ed0000 MEMATTR=0x1390b8118 ACPI 2.0=0x136760018 RNG=0x13676e918 MEMRESERVE=0x136b41218 Oct 13 00:01:33.830731 kernel: random: crng init done Oct 13 00:01:33.830738 kernel: secureboot: Secure boot disabled Oct 13 00:01:33.830743 kernel: ACPI: Early table checksum verification disabled Oct 13 00:01:33.830749 kernel: ACPI: RSDP 0x0000000136760018 000024 (v02 BOCHS ) Oct 13 00:01:33.830754 kernel: ACPI: XSDT 0x000000013676FE98 00006C (v01 BOCHS BXPC 00000001 01000013) Oct 13 00:01:33.830762 kernel: ACPI: FACP 0x000000013676FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830767 kernel: ACPI: DSDT 0x0000000136767518 001468 (v02 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830773 kernel: ACPI: APIC 0x000000013676FC18 000108 (v04 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830779 kernel: ACPI: PPTT 0x000000013676FD98 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830786 kernel: ACPI: GTDT 0x000000013676D898 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830793 kernel: ACPI: MCFG 0x000000013676FF98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830799 kernel: ACPI: SPCR 0x000000013676E818 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830805 kernel: ACPI: DBG2 0x000000013676E898 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830811 kernel: ACPI: IORT 0x000000013676E418 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Oct 13 00:01:33.830817 kernel: ACPI: BGRT 0x000000013676E798 000038 (v01 INTEL EDK2 00000002 01000013) Oct 13 00:01:33.830823 kernel: ACPI: SPCR: console: pl011,mmio32,0x9000000,9600 Oct 13 00:01:33.830829 kernel: ACPI: Use ACPI SPCR as default console: No Oct 13 00:01:33.830835 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x0000000139ffffff] Oct 13 00:01:33.830841 kernel: NODE_DATA(0) allocated [mem 0x13967da00-0x139684fff] Oct 13 00:01:33.830847 kernel: Zone ranges: Oct 13 00:01:33.830853 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Oct 13 00:01:33.830860 kernel: DMA32 empty Oct 13 00:01:33.830866 kernel: Normal [mem 0x0000000100000000-0x0000000139ffffff] Oct 13 00:01:33.830872 kernel: Device empty Oct 13 00:01:33.830878 kernel: Movable zone start for each node Oct 13 00:01:33.830884 kernel: Early memory node ranges Oct 13 00:01:33.830890 kernel: node 0: [mem 0x0000000040000000-0x000000013666ffff] Oct 13 00:01:33.830895 kernel: node 0: [mem 0x0000000136670000-0x000000013667ffff] Oct 13 00:01:33.830901 kernel: node 0: [mem 0x0000000136680000-0x000000013676ffff] Oct 13 00:01:33.830907 kernel: node 0: [mem 0x0000000136770000-0x0000000136b3ffff] Oct 13 00:01:33.830913 kernel: node 0: [mem 0x0000000136b40000-0x0000000139e1ffff] Oct 13 00:01:33.830919 kernel: node 0: [mem 0x0000000139e20000-0x0000000139eaffff] Oct 13 00:01:33.830925 kernel: node 0: [mem 0x0000000139eb0000-0x0000000139ebffff] Oct 13 00:01:33.830933 kernel: node 0: [mem 0x0000000139ec0000-0x0000000139fdffff] Oct 13 00:01:33.830939 kernel: node 0: [mem 0x0000000139fe0000-0x0000000139ffffff] Oct 13 00:01:33.830965 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x0000000139ffffff] Oct 13 00:01:33.830973 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Oct 13 00:01:33.830979 kernel: cma: Reserved 16 MiB at 0x00000000ff000000 on node -1 Oct 13 00:01:33.830987 kernel: psci: probing for conduit method from ACPI. Oct 13 00:01:33.830993 kernel: psci: PSCIv1.1 detected in firmware. Oct 13 00:01:33.831000 kernel: psci: Using standard PSCI v0.2 function IDs Oct 13 00:01:33.831006 kernel: psci: Trusted OS migration not required Oct 13 00:01:33.831012 kernel: psci: SMC Calling Convention v1.1 Oct 13 00:01:33.831019 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Oct 13 00:01:33.831025 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Oct 13 00:01:33.831044 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Oct 13 00:01:33.831052 kernel: pcpu-alloc: [0] 0 [0] 1 Oct 13 00:01:33.831058 kernel: Detected PIPT I-cache on CPU0 Oct 13 00:01:33.831064 kernel: CPU features: detected: GIC system register CPU interface Oct 13 00:01:33.831073 kernel: CPU features: detected: Spectre-v4 Oct 13 00:01:33.831079 kernel: CPU features: detected: Spectre-BHB Oct 13 00:01:33.831086 kernel: CPU features: kernel page table isolation forced ON by KASLR Oct 13 00:01:33.831092 kernel: CPU features: detected: Kernel page table isolation (KPTI) Oct 13 00:01:33.831098 kernel: CPU features: detected: ARM erratum 1418040 Oct 13 00:01:33.831105 kernel: CPU features: detected: SSBS not fully self-synchronizing Oct 13 00:01:33.831111 kernel: alternatives: applying boot alternatives Oct 13 00:01:33.831119 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=37fc523060a9b8894388e25ab0f082059dd744d472a2b8577211d4b3dd66a910 Oct 13 00:01:33.831126 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Oct 13 00:01:33.831132 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Oct 13 00:01:33.831144 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Oct 13 00:01:33.831151 kernel: Fallback order for Node 0: 0 Oct 13 00:01:33.831158 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1024000 Oct 13 00:01:33.831164 kernel: Policy zone: Normal Oct 13 00:01:33.831171 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Oct 13 00:01:33.831177 kernel: software IO TLB: area num 2. Oct 13 00:01:33.831184 kernel: software IO TLB: mapped [mem 0x00000000fb000000-0x00000000ff000000] (64MB) Oct 13 00:01:33.831190 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Oct 13 00:01:33.831197 kernel: rcu: Preemptible hierarchical RCU implementation. Oct 13 00:01:33.831204 kernel: rcu: RCU event tracing is enabled. Oct 13 00:01:33.831211 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Oct 13 00:01:33.831217 kernel: Trampoline variant of Tasks RCU enabled. Oct 13 00:01:33.831226 kernel: Tracing variant of Tasks RCU enabled. Oct 13 00:01:33.831232 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Oct 13 00:01:33.831239 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Oct 13 00:01:33.831245 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 13 00:01:33.831252 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 13 00:01:33.831259 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Oct 13 00:01:33.831265 kernel: GICv3: 256 SPIs implemented Oct 13 00:01:33.831272 kernel: GICv3: 0 Extended SPIs implemented Oct 13 00:01:33.831278 kernel: Root IRQ handler: gic_handle_irq Oct 13 00:01:33.831285 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Oct 13 00:01:33.831291 kernel: GICv3: GICD_CTRL.DS=1, SCR_EL3.FIQ=0 Oct 13 00:01:33.831298 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Oct 13 00:01:33.831306 kernel: ITS [mem 0x08080000-0x0809ffff] Oct 13 00:01:33.831312 kernel: ITS@0x0000000008080000: allocated 8192 Devices @100100000 (indirect, esz 8, psz 64K, shr 1) Oct 13 00:01:33.831319 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @100110000 (flat, esz 8, psz 64K, shr 1) Oct 13 00:01:33.831325 kernel: GICv3: using LPI property table @0x0000000100120000 Oct 13 00:01:33.831332 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000100130000 Oct 13 00:01:33.831339 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Oct 13 00:01:33.831345 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Oct 13 00:01:33.831354 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Oct 13 00:01:33.831361 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Oct 13 00:01:33.831368 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Oct 13 00:01:33.831374 kernel: Console: colour dummy device 80x25 Oct 13 00:01:33.831385 kernel: ACPI: Core revision 20240827 Oct 13 00:01:33.831392 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Oct 13 00:01:33.831398 kernel: pid_max: default: 32768 minimum: 301 Oct 13 00:01:33.831422 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Oct 13 00:01:33.831429 kernel: landlock: Up and running. Oct 13 00:01:33.831436 kernel: SELinux: Initializing. Oct 13 00:01:33.831442 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Oct 13 00:01:33.831451 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Oct 13 00:01:33.831458 kernel: rcu: Hierarchical SRCU implementation. Oct 13 00:01:33.831467 kernel: rcu: Max phase no-delay instances is 400. Oct 13 00:01:33.831477 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Oct 13 00:01:33.831484 kernel: Remapping and enabling EFI services. Oct 13 00:01:33.831491 kernel: smp: Bringing up secondary CPUs ... Oct 13 00:01:33.831498 kernel: Detected PIPT I-cache on CPU1 Oct 13 00:01:33.831507 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Oct 13 00:01:33.831514 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000100140000 Oct 13 00:01:33.831522 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Oct 13 00:01:33.831530 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Oct 13 00:01:33.831538 kernel: smp: Brought up 1 node, 2 CPUs Oct 13 00:01:33.831552 kernel: SMP: Total of 2 processors activated. Oct 13 00:01:33.831558 kernel: CPU: All CPU(s) started at EL1 Oct 13 00:01:33.831567 kernel: CPU features: detected: 32-bit EL0 Support Oct 13 00:01:33.831574 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Oct 13 00:01:33.831581 kernel: CPU features: detected: Common not Private translations Oct 13 00:01:33.831588 kernel: CPU features: detected: CRC32 instructions Oct 13 00:01:33.831595 kernel: CPU features: detected: Enhanced Virtualization Traps Oct 13 00:01:33.831603 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Oct 13 00:01:33.831610 kernel: CPU features: detected: LSE atomic instructions Oct 13 00:01:33.831617 kernel: CPU features: detected: Privileged Access Never Oct 13 00:01:33.831624 kernel: CPU features: detected: RAS Extension Support Oct 13 00:01:33.831631 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Oct 13 00:01:33.831638 kernel: alternatives: applying system-wide alternatives Oct 13 00:01:33.831645 kernel: CPU features: detected: Hardware dirty bit management on CPU0-1 Oct 13 00:01:33.831652 kernel: Memory: 3859556K/4096000K available (11136K kernel code, 2450K rwdata, 9076K rodata, 38976K init, 1038K bss, 214964K reserved, 16384K cma-reserved) Oct 13 00:01:33.831659 kernel: devtmpfs: initialized Oct 13 00:01:33.831667 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Oct 13 00:01:33.831674 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Oct 13 00:01:33.831681 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Oct 13 00:01:33.831688 kernel: 0 pages in range for non-PLT usage Oct 13 00:01:33.831695 kernel: 508560 pages in range for PLT usage Oct 13 00:01:33.831702 kernel: pinctrl core: initialized pinctrl subsystem Oct 13 00:01:33.831708 kernel: SMBIOS 3.0.0 present. Oct 13 00:01:33.831715 kernel: DMI: Hetzner vServer/KVM Virtual Machine, BIOS 20171111 11/11/2017 Oct 13 00:01:33.831722 kernel: DMI: Memory slots populated: 1/1 Oct 13 00:01:33.831730 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Oct 13 00:01:33.831737 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Oct 13 00:01:33.831744 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Oct 13 00:01:33.831751 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Oct 13 00:01:33.831758 kernel: audit: initializing netlink subsys (disabled) Oct 13 00:01:33.831765 kernel: audit: type=2000 audit(0.018:1): state=initialized audit_enabled=0 res=1 Oct 13 00:01:33.831772 kernel: thermal_sys: Registered thermal governor 'step_wise' Oct 13 00:01:33.831779 kernel: cpuidle: using governor menu Oct 13 00:01:33.831786 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Oct 13 00:01:33.831794 kernel: ASID allocator initialised with 32768 entries Oct 13 00:01:33.831801 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Oct 13 00:01:33.831808 kernel: Serial: AMBA PL011 UART driver Oct 13 00:01:33.831815 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Oct 13 00:01:33.831824 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Oct 13 00:01:33.831833 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Oct 13 00:01:33.831841 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Oct 13 00:01:33.831849 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Oct 13 00:01:33.831856 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Oct 13 00:01:33.831866 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Oct 13 00:01:33.831874 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Oct 13 00:01:33.831881 kernel: ACPI: Added _OSI(Module Device) Oct 13 00:01:33.831888 kernel: ACPI: Added _OSI(Processor Device) Oct 13 00:01:33.831895 kernel: ACPI: Added _OSI(Processor Aggregator Device) Oct 13 00:01:33.831902 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Oct 13 00:01:33.831909 kernel: ACPI: Interpreter enabled Oct 13 00:01:33.831915 kernel: ACPI: Using GIC for interrupt routing Oct 13 00:01:33.831922 kernel: ACPI: MCFG table detected, 1 entries Oct 13 00:01:33.831930 kernel: ACPI: CPU0 has been hot-added Oct 13 00:01:33.831937 kernel: ACPI: CPU1 has been hot-added Oct 13 00:01:33.831954 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Oct 13 00:01:33.831961 kernel: printk: legacy console [ttyAMA0] enabled Oct 13 00:01:33.831968 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Oct 13 00:01:33.832124 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Oct 13 00:01:33.832189 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Oct 13 00:01:33.832271 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Oct 13 00:01:33.832345 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Oct 13 00:01:33.832849 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Oct 13 00:01:33.832869 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Oct 13 00:01:33.832879 kernel: PCI host bridge to bus 0000:00 Oct 13 00:01:33.833026 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Oct 13 00:01:33.833098 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Oct 13 00:01:33.833155 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Oct 13 00:01:33.833235 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Oct 13 00:01:33.833330 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 conventional PCI endpoint Oct 13 00:01:33.833507 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x038000 conventional PCI endpoint Oct 13 00:01:33.833600 kernel: pci 0000:00:01.0: BAR 1 [mem 0x11289000-0x11289fff] Oct 13 00:01:33.833678 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000600000-0x8000603fff 64bit pref] Oct 13 00:01:33.833757 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.833821 kernel: pci 0000:00:02.0: BAR 0 [mem 0x11288000-0x11288fff] Oct 13 00:01:33.833880 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Oct 13 00:01:33.833938 kernel: pci 0000:00:02.0: bridge window [mem 0x11000000-0x111fffff] Oct 13 00:01:33.834014 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80000fffff 64bit pref] Oct 13 00:01:33.834092 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.834158 kernel: pci 0000:00:02.1: BAR 0 [mem 0x11287000-0x11287fff] Oct 13 00:01:33.834222 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Oct 13 00:01:33.834294 kernel: pci 0000:00:02.1: bridge window [mem 0x10e00000-0x10ffffff] Oct 13 00:01:33.834363 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.834440 kernel: pci 0000:00:02.2: BAR 0 [mem 0x11286000-0x11286fff] Oct 13 00:01:33.834518 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Oct 13 00:01:33.834583 kernel: pci 0000:00:02.2: bridge window [mem 0x10c00000-0x10dfffff] Oct 13 00:01:33.834653 kernel: pci 0000:00:02.2: bridge window [mem 0x8000100000-0x80001fffff 64bit pref] Oct 13 00:01:33.834734 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.834813 kernel: pci 0000:00:02.3: BAR 0 [mem 0x11285000-0x11285fff] Oct 13 00:01:33.834873 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Oct 13 00:01:33.834931 kernel: pci 0000:00:02.3: bridge window [mem 0x10a00000-0x10bfffff] Oct 13 00:01:33.835002 kernel: pci 0000:00:02.3: bridge window [mem 0x8000200000-0x80002fffff 64bit pref] Oct 13 00:01:33.835086 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.835161 kernel: pci 0000:00:02.4: BAR 0 [mem 0x11284000-0x11284fff] Oct 13 00:01:33.835234 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Oct 13 00:01:33.835311 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Oct 13 00:01:33.835401 kernel: pci 0000:00:02.4: bridge window [mem 0x8000300000-0x80003fffff 64bit pref] Oct 13 00:01:33.835545 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.835610 kernel: pci 0000:00:02.5: BAR 0 [mem 0x11283000-0x11283fff] Oct 13 00:01:33.835675 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Oct 13 00:01:33.835740 kernel: pci 0000:00:02.5: bridge window [mem 0x10600000-0x107fffff] Oct 13 00:01:33.835804 kernel: pci 0000:00:02.5: bridge window [mem 0x8000400000-0x80004fffff 64bit pref] Oct 13 00:01:33.835884 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.835990 kernel: pci 0000:00:02.6: BAR 0 [mem 0x11282000-0x11282fff] Oct 13 00:01:33.836072 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Oct 13 00:01:33.836133 kernel: pci 0000:00:02.6: bridge window [mem 0x10400000-0x105fffff] Oct 13 00:01:33.836192 kernel: pci 0000:00:02.6: bridge window [mem 0x8000500000-0x80005fffff 64bit pref] Oct 13 00:01:33.836259 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.836318 kernel: pci 0000:00:02.7: BAR 0 [mem 0x11281000-0x11281fff] Oct 13 00:01:33.836380 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Oct 13 00:01:33.836458 kernel: pci 0000:00:02.7: bridge window [mem 0x10200000-0x103fffff] Oct 13 00:01:33.836529 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Oct 13 00:01:33.836588 kernel: pci 0000:00:03.0: BAR 0 [mem 0x11280000-0x11280fff] Oct 13 00:01:33.836645 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Oct 13 00:01:33.836703 kernel: pci 0000:00:03.0: bridge window [mem 0x10000000-0x101fffff] Oct 13 00:01:33.836776 kernel: pci 0000:00:04.0: [1b36:0002] type 00 class 0x070002 conventional PCI endpoint Oct 13 00:01:33.836834 kernel: pci 0000:00:04.0: BAR 0 [io 0x0000-0x0007] Oct 13 00:01:33.836903 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Oct 13 00:01:33.836979 kernel: pci 0000:01:00.0: BAR 1 [mem 0x11000000-0x11000fff] Oct 13 00:01:33.837041 kernel: pci 0000:01:00.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref] Oct 13 00:01:33.837101 kernel: pci 0000:01:00.0: ROM [mem 0xfff80000-0xffffffff pref] Oct 13 00:01:33.837174 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 PCIe Endpoint Oct 13 00:01:33.837239 kernel: pci 0000:02:00.0: BAR 0 [mem 0x10e00000-0x10e03fff 64bit] Oct 13 00:01:33.837315 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 PCIe Endpoint Oct 13 00:01:33.837375 kernel: pci 0000:03:00.0: BAR 1 [mem 0x10c00000-0x10c00fff] Oct 13 00:01:33.837459 kernel: pci 0000:03:00.0: BAR 4 [mem 0x8000100000-0x8000103fff 64bit pref] Oct 13 00:01:33.837529 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 PCIe Endpoint Oct 13 00:01:33.837591 kernel: pci 0000:04:00.0: BAR 4 [mem 0x8000200000-0x8000203fff 64bit pref] Oct 13 00:01:33.837670 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 PCIe Endpoint Oct 13 00:01:33.837744 kernel: pci 0000:05:00.0: BAR 1 [mem 0x10800000-0x10800fff] Oct 13 00:01:33.837816 kernel: pci 0000:05:00.0: BAR 4 [mem 0x8000300000-0x8000303fff 64bit pref] Oct 13 00:01:33.837897 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 PCIe Endpoint Oct 13 00:01:33.837970 kernel: pci 0000:06:00.0: BAR 1 [mem 0x10600000-0x10600fff] Oct 13 00:01:33.838036 kernel: pci 0000:06:00.0: BAR 4 [mem 0x8000400000-0x8000403fff 64bit pref] Oct 13 00:01:33.838114 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Oct 13 00:01:33.838193 kernel: pci 0000:07:00.0: BAR 1 [mem 0x10400000-0x10400fff] Oct 13 00:01:33.838303 kernel: pci 0000:07:00.0: BAR 4 [mem 0x8000500000-0x8000503fff 64bit pref] Oct 13 00:01:33.838453 kernel: pci 0000:07:00.0: ROM [mem 0xfff80000-0xffffffff pref] Oct 13 00:01:33.838551 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Oct 13 00:01:33.838615 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 01] add_size 100000 add_align 100000 Oct 13 00:01:33.838676 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff] to [bus 01] add_size 100000 add_align 100000 Oct 13 00:01:33.838740 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Oct 13 00:01:33.838808 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Oct 13 00:01:33.838873 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x001fffff] to [bus 02] add_size 100000 add_align 100000 Oct 13 00:01:33.838937 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Oct 13 00:01:33.839031 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 03] add_size 100000 add_align 100000 Oct 13 00:01:33.839121 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Oct 13 00:01:33.839195 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Oct 13 00:01:33.839263 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 04] add_size 100000 add_align 100000 Oct 13 00:01:33.839326 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Oct 13 00:01:33.839389 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Oct 13 00:01:33.839675 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 05] add_size 100000 add_align 100000 Oct 13 00:01:33.839750 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff] to [bus 05] add_size 100000 add_align 100000 Oct 13 00:01:33.839814 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Oct 13 00:01:33.839875 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 06] add_size 100000 add_align 100000 Oct 13 00:01:33.839940 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff] to [bus 06] add_size 100000 add_align 100000 Oct 13 00:01:33.840071 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Oct 13 00:01:33.840134 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 07] add_size 100000 add_align 100000 Oct 13 00:01:33.840193 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff] to [bus 07] add_size 100000 add_align 100000 Oct 13 00:01:33.840258 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Oct 13 00:01:33.840318 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 08] add_size 200000 add_align 100000 Oct 13 00:01:33.840376 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff] to [bus 08] add_size 200000 add_align 100000 Oct 13 00:01:33.840494 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Oct 13 00:01:33.840555 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 09] add_size 200000 add_align 100000 Oct 13 00:01:33.840614 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 09] add_size 200000 add_align 100000 Oct 13 00:01:33.840674 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff]: assigned Oct 13 00:01:33.840733 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref]: assigned Oct 13 00:01:33.840795 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff]: assigned Oct 13 00:01:33.840853 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref]: assigned Oct 13 00:01:33.840919 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff]: assigned Oct 13 00:01:33.840995 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref]: assigned Oct 13 00:01:33.841066 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff]: assigned Oct 13 00:01:33.841127 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref]: assigned Oct 13 00:01:33.841188 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff]: assigned Oct 13 00:01:33.841262 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref]: assigned Oct 13 00:01:33.841337 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff]: assigned Oct 13 00:01:33.841398 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref]: assigned Oct 13 00:01:33.842230 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff]: assigned Oct 13 00:01:33.842331 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref]: assigned Oct 13 00:01:33.842553 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff]: assigned Oct 13 00:01:33.842637 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref]: assigned Oct 13 00:01:33.842704 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff]: assigned Oct 13 00:01:33.842765 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref]: assigned Oct 13 00:01:33.842846 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8001200000-0x8001203fff 64bit pref]: assigned Oct 13 00:01:33.842912 kernel: pci 0000:00:01.0: BAR 1 [mem 0x11200000-0x11200fff]: assigned Oct 13 00:01:33.843069 kernel: pci 0000:00:02.0: BAR 0 [mem 0x11201000-0x11201fff]: assigned Oct 13 00:01:33.843153 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff]: assigned Oct 13 00:01:33.843225 kernel: pci 0000:00:02.1: BAR 0 [mem 0x11202000-0x11202fff]: assigned Oct 13 00:01:33.843286 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff]: assigned Oct 13 00:01:33.843356 kernel: pci 0000:00:02.2: BAR 0 [mem 0x11203000-0x11203fff]: assigned Oct 13 00:01:33.843456 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff]: assigned Oct 13 00:01:33.843524 kernel: pci 0000:00:02.3: BAR 0 [mem 0x11204000-0x11204fff]: assigned Oct 13 00:01:33.843600 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff]: assigned Oct 13 00:01:33.843677 kernel: pci 0000:00:02.4: BAR 0 [mem 0x11205000-0x11205fff]: assigned Oct 13 00:01:33.843747 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff]: assigned Oct 13 00:01:33.843824 kernel: pci 0000:00:02.5: BAR 0 [mem 0x11206000-0x11206fff]: assigned Oct 13 00:01:33.843895 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff]: assigned Oct 13 00:01:33.843994 kernel: pci 0000:00:02.6: BAR 0 [mem 0x11207000-0x11207fff]: assigned Oct 13 00:01:33.844072 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff]: assigned Oct 13 00:01:33.844146 kernel: pci 0000:00:02.7: BAR 0 [mem 0x11208000-0x11208fff]: assigned Oct 13 00:01:33.844219 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff]: assigned Oct 13 00:01:33.844294 kernel: pci 0000:00:03.0: BAR 0 [mem 0x11209000-0x11209fff]: assigned Oct 13 00:01:33.844359 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff]: assigned Oct 13 00:01:33.846322 kernel: pci 0000:00:04.0: BAR 0 [io 0xa000-0xa007]: assigned Oct 13 00:01:33.846494 kernel: pci 0000:01:00.0: ROM [mem 0x10000000-0x1007ffff pref]: assigned Oct 13 00:01:33.846576 kernel: pci 0000:01:00.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref]: assigned Oct 13 00:01:33.846639 kernel: pci 0000:01:00.0: BAR 1 [mem 0x10080000-0x10080fff]: assigned Oct 13 00:01:33.846708 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Oct 13 00:01:33.846771 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Oct 13 00:01:33.846830 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff] Oct 13 00:01:33.846893 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref] Oct 13 00:01:33.846979 kernel: pci 0000:02:00.0: BAR 0 [mem 0x10200000-0x10203fff 64bit]: assigned Oct 13 00:01:33.847049 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Oct 13 00:01:33.847109 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Oct 13 00:01:33.847166 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff] Oct 13 00:01:33.847224 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref] Oct 13 00:01:33.847289 kernel: pci 0000:03:00.0: BAR 4 [mem 0x8000400000-0x8000403fff 64bit pref]: assigned Oct 13 00:01:33.847350 kernel: pci 0000:03:00.0: BAR 1 [mem 0x10400000-0x10400fff]: assigned Oct 13 00:01:33.847458 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Oct 13 00:01:33.847530 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Oct 13 00:01:33.847597 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff] Oct 13 00:01:33.847656 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref] Oct 13 00:01:33.847727 kernel: pci 0000:04:00.0: BAR 4 [mem 0x8000600000-0x8000603fff 64bit pref]: assigned Oct 13 00:01:33.847788 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Oct 13 00:01:33.847850 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Oct 13 00:01:33.847912 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff] Oct 13 00:01:33.847988 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref] Oct 13 00:01:33.848060 kernel: pci 0000:05:00.0: BAR 4 [mem 0x8000800000-0x8000803fff 64bit pref]: assigned Oct 13 00:01:33.848127 kernel: pci 0000:05:00.0: BAR 1 [mem 0x10800000-0x10800fff]: assigned Oct 13 00:01:33.848188 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Oct 13 00:01:33.848246 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Oct 13 00:01:33.848304 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Oct 13 00:01:33.848361 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref] Oct 13 00:01:33.848453 kernel: pci 0000:06:00.0: BAR 4 [mem 0x8000a00000-0x8000a03fff 64bit pref]: assigned Oct 13 00:01:33.849460 kernel: pci 0000:06:00.0: BAR 1 [mem 0x10a00000-0x10a00fff]: assigned Oct 13 00:01:33.849608 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Oct 13 00:01:33.849696 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Oct 13 00:01:33.849764 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff] Oct 13 00:01:33.849833 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref] Oct 13 00:01:33.849907 kernel: pci 0000:07:00.0: ROM [mem 0x10c00000-0x10c7ffff pref]: assigned Oct 13 00:01:33.849988 kernel: pci 0000:07:00.0: BAR 4 [mem 0x8000c00000-0x8000c03fff 64bit pref]: assigned Oct 13 00:01:33.850097 kernel: pci 0000:07:00.0: BAR 1 [mem 0x10c80000-0x10c80fff]: assigned Oct 13 00:01:33.850165 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Oct 13 00:01:33.850228 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Oct 13 00:01:33.850289 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff] Oct 13 00:01:33.850349 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref] Oct 13 00:01:33.850432 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Oct 13 00:01:33.850496 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Oct 13 00:01:33.850555 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff] Oct 13 00:01:33.850615 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref] Oct 13 00:01:33.850677 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Oct 13 00:01:33.850745 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff] Oct 13 00:01:33.850803 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff] Oct 13 00:01:33.850862 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref] Oct 13 00:01:33.850926 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Oct 13 00:01:33.850993 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Oct 13 00:01:33.851047 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Oct 13 00:01:33.851115 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Oct 13 00:01:33.851174 kernel: pci_bus 0000:01: resource 1 [mem 0x10000000-0x101fffff] Oct 13 00:01:33.851228 kernel: pci_bus 0000:01: resource 2 [mem 0x8000000000-0x80001fffff 64bit pref] Oct 13 00:01:33.851293 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x2fff] Oct 13 00:01:33.851348 kernel: pci_bus 0000:02: resource 1 [mem 0x10200000-0x103fffff] Oct 13 00:01:33.851428 kernel: pci_bus 0000:02: resource 2 [mem 0x8000200000-0x80003fffff 64bit pref] Oct 13 00:01:33.851502 kernel: pci_bus 0000:03: resource 0 [io 0x3000-0x3fff] Oct 13 00:01:33.851568 kernel: pci_bus 0000:03: resource 1 [mem 0x10400000-0x105fffff] Oct 13 00:01:33.851622 kernel: pci_bus 0000:03: resource 2 [mem 0x8000400000-0x80005fffff 64bit pref] Oct 13 00:01:33.851699 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Oct 13 00:01:33.851766 kernel: pci_bus 0000:04: resource 1 [mem 0x10600000-0x107fffff] Oct 13 00:01:33.851835 kernel: pci_bus 0000:04: resource 2 [mem 0x8000600000-0x80007fffff 64bit pref] Oct 13 00:01:33.851911 kernel: pci_bus 0000:05: resource 0 [io 0x5000-0x5fff] Oct 13 00:01:33.852018 kernel: pci_bus 0000:05: resource 1 [mem 0x10800000-0x109fffff] Oct 13 00:01:33.852086 kernel: pci_bus 0000:05: resource 2 [mem 0x8000800000-0x80009fffff 64bit pref] Oct 13 00:01:33.852168 kernel: pci_bus 0000:06: resource 0 [io 0x6000-0x6fff] Oct 13 00:01:33.852243 kernel: pci_bus 0000:06: resource 1 [mem 0x10a00000-0x10bfffff] Oct 13 00:01:33.852301 kernel: pci_bus 0000:06: resource 2 [mem 0x8000a00000-0x8000bfffff 64bit pref] Oct 13 00:01:33.852368 kernel: pci_bus 0000:07: resource 0 [io 0x7000-0x7fff] Oct 13 00:01:33.854817 kernel: pci_bus 0000:07: resource 1 [mem 0x10c00000-0x10dfffff] Oct 13 00:01:33.854906 kernel: pci_bus 0000:07: resource 2 [mem 0x8000c00000-0x8000dfffff 64bit pref] Oct 13 00:01:33.856576 kernel: pci_bus 0000:08: resource 0 [io 0x8000-0x8fff] Oct 13 00:01:33.856673 kernel: pci_bus 0000:08: resource 1 [mem 0x10e00000-0x10ffffff] Oct 13 00:01:33.856747 kernel: pci_bus 0000:08: resource 2 [mem 0x8000e00000-0x8000ffffff 64bit pref] Oct 13 00:01:33.856835 kernel: pci_bus 0000:09: resource 0 [io 0x9000-0x9fff] Oct 13 00:01:33.856894 kernel: pci_bus 0000:09: resource 1 [mem 0x11000000-0x111fffff] Oct 13 00:01:33.856967 kernel: pci_bus 0000:09: resource 2 [mem 0x8001000000-0x80011fffff 64bit pref] Oct 13 00:01:33.856986 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Oct 13 00:01:33.856994 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Oct 13 00:01:33.857001 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Oct 13 00:01:33.857009 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Oct 13 00:01:33.857016 kernel: iommu: Default domain type: Translated Oct 13 00:01:33.857024 kernel: iommu: DMA domain TLB invalidation policy: strict mode Oct 13 00:01:33.857031 kernel: efivars: Registered efivars operations Oct 13 00:01:33.857038 kernel: vgaarb: loaded Oct 13 00:01:33.857046 kernel: clocksource: Switched to clocksource arch_sys_counter Oct 13 00:01:33.857053 kernel: VFS: Disk quotas dquot_6.6.0 Oct 13 00:01:33.857063 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Oct 13 00:01:33.857070 kernel: pnp: PnP ACPI init Oct 13 00:01:33.857146 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Oct 13 00:01:33.857159 kernel: pnp: PnP ACPI: found 1 devices Oct 13 00:01:33.857168 kernel: NET: Registered PF_INET protocol family Oct 13 00:01:33.857176 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Oct 13 00:01:33.857185 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Oct 13 00:01:33.857193 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Oct 13 00:01:33.857203 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Oct 13 00:01:33.857210 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Oct 13 00:01:33.857219 kernel: TCP: Hash tables configured (established 32768 bind 32768) Oct 13 00:01:33.857226 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Oct 13 00:01:33.857234 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Oct 13 00:01:33.857244 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Oct 13 00:01:33.857326 kernel: pci 0000:02:00.0: enabling device (0000 -> 0002) Oct 13 00:01:33.857339 kernel: PCI: CLS 0 bytes, default 64 Oct 13 00:01:33.857346 kernel: kvm [1]: HYP mode not available Oct 13 00:01:33.857355 kernel: Initialise system trusted keyrings Oct 13 00:01:33.857363 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Oct 13 00:01:33.857370 kernel: Key type asymmetric registered Oct 13 00:01:33.857377 kernel: Asymmetric key parser 'x509' registered Oct 13 00:01:33.857387 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Oct 13 00:01:33.857394 kernel: io scheduler mq-deadline registered Oct 13 00:01:33.857401 kernel: io scheduler kyber registered Oct 13 00:01:33.857463 kernel: io scheduler bfq registered Oct 13 00:01:33.857471 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Oct 13 00:01:33.857558 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 50 Oct 13 00:01:33.857632 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 50 Oct 13 00:01:33.857693 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.857756 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 51 Oct 13 00:01:33.857817 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 51 Oct 13 00:01:33.857876 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.857940 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 52 Oct 13 00:01:33.858059 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 52 Oct 13 00:01:33.858124 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.858188 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 53 Oct 13 00:01:33.858250 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 53 Oct 13 00:01:33.858310 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.858372 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 54 Oct 13 00:01:33.859598 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 54 Oct 13 00:01:33.859681 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.859752 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 55 Oct 13 00:01:33.859812 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 55 Oct 13 00:01:33.859871 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.860031 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 56 Oct 13 00:01:33.860119 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 56 Oct 13 00:01:33.860190 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.860270 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 57 Oct 13 00:01:33.860331 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 57 Oct 13 00:01:33.860400 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.861472 kernel: ACPI: \_SB_.PCI0.GSI3: Enabled at IRQ 38 Oct 13 00:01:33.861599 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 58 Oct 13 00:01:33.861671 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 58 Oct 13 00:01:33.861737 kernel: pcieport 0000:00:03.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 13 00:01:33.861749 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Oct 13 00:01:33.861757 kernel: ACPI: button: Power Button [PWRB] Oct 13 00:01:33.861765 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Oct 13 00:01:33.861838 kernel: virtio-pci 0000:04:00.0: enabling device (0000 -> 0002) Oct 13 00:01:33.861912 kernel: virtio-pci 0000:07:00.0: enabling device (0000 -> 0002) Oct 13 00:01:33.861923 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Oct 13 00:01:33.861931 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Oct 13 00:01:33.862015 kernel: serial 0000:00:04.0: enabling device (0000 -> 0001) Oct 13 00:01:33.862027 kernel: 0000:00:04.0: ttyS0 at I/O 0xa000 (irq = 45, base_baud = 115200) is a 16550A Oct 13 00:01:33.862035 kernel: thunder_xcv, ver 1.0 Oct 13 00:01:33.862043 kernel: thunder_bgx, ver 1.0 Oct 13 00:01:33.862051 kernel: nicpf, ver 1.0 Oct 13 00:01:33.862061 kernel: nicvf, ver 1.0 Oct 13 00:01:33.862143 kernel: rtc-efi rtc-efi.0: registered as rtc0 Oct 13 00:01:33.862202 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-10-13T00:01:33 UTC (1760313693) Oct 13 00:01:33.862212 kernel: hid: raw HID events driver (C) Jiri Kosina Oct 13 00:01:33.862223 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Oct 13 00:01:33.862232 kernel: watchdog: NMI not fully supported Oct 13 00:01:33.862241 kernel: watchdog: Hard watchdog permanently disabled Oct 13 00:01:33.862249 kernel: NET: Registered PF_INET6 protocol family Oct 13 00:01:33.862260 kernel: Segment Routing with IPv6 Oct 13 00:01:33.862269 kernel: In-situ OAM (IOAM) with IPv6 Oct 13 00:01:33.862277 kernel: NET: Registered PF_PACKET protocol family Oct 13 00:01:33.862285 kernel: Key type dns_resolver registered Oct 13 00:01:33.862293 kernel: registered taskstats version 1 Oct 13 00:01:33.862300 kernel: Loading compiled-in X.509 certificates Oct 13 00:01:33.862308 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.51-flatcar: b8447a1087a9e9c4d5b9d4c2f2bba5a69a74f139' Oct 13 00:01:33.862315 kernel: Demotion targets for Node 0: null Oct 13 00:01:33.862323 kernel: Key type .fscrypt registered Oct 13 00:01:33.862331 kernel: Key type fscrypt-provisioning registered Oct 13 00:01:33.862339 kernel: ima: No TPM chip found, activating TPM-bypass! Oct 13 00:01:33.862346 kernel: ima: Allocated hash algorithm: sha1 Oct 13 00:01:33.862354 kernel: ima: No architecture policies found Oct 13 00:01:33.862361 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Oct 13 00:01:33.862369 kernel: clk: Disabling unused clocks Oct 13 00:01:33.862376 kernel: PM: genpd: Disabling unused power domains Oct 13 00:01:33.862383 kernel: Warning: unable to open an initial console. Oct 13 00:01:33.862391 kernel: Freeing unused kernel memory: 38976K Oct 13 00:01:33.862400 kernel: Run /init as init process Oct 13 00:01:33.862443 kernel: with arguments: Oct 13 00:01:33.862454 kernel: /init Oct 13 00:01:33.862461 kernel: with environment: Oct 13 00:01:33.862468 kernel: HOME=/ Oct 13 00:01:33.862476 kernel: TERM=linux Oct 13 00:01:33.862483 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Oct 13 00:01:33.862492 systemd[1]: Successfully made /usr/ read-only. Oct 13 00:01:33.862504 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 13 00:01:33.862513 systemd[1]: Detected virtualization kvm. Oct 13 00:01:33.862520 systemd[1]: Detected architecture arm64. Oct 13 00:01:33.862528 systemd[1]: Running in initrd. Oct 13 00:01:33.862535 systemd[1]: No hostname configured, using default hostname. Oct 13 00:01:33.862543 systemd[1]: Hostname set to . Oct 13 00:01:33.862551 systemd[1]: Initializing machine ID from VM UUID. Oct 13 00:01:33.862559 systemd[1]: Queued start job for default target initrd.target. Oct 13 00:01:33.862568 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 13 00:01:33.862576 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 13 00:01:33.862584 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Oct 13 00:01:33.862592 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 13 00:01:33.862600 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Oct 13 00:01:33.862612 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Oct 13 00:01:33.862622 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Oct 13 00:01:33.862634 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Oct 13 00:01:33.862642 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 13 00:01:33.862650 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 13 00:01:33.862658 systemd[1]: Reached target paths.target - Path Units. Oct 13 00:01:33.862666 systemd[1]: Reached target slices.target - Slice Units. Oct 13 00:01:33.862674 systemd[1]: Reached target swap.target - Swaps. Oct 13 00:01:33.862681 systemd[1]: Reached target timers.target - Timer Units. Oct 13 00:01:33.862689 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Oct 13 00:01:33.862698 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 13 00:01:33.862706 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 13 00:01:33.862714 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Oct 13 00:01:33.862723 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 13 00:01:33.862731 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 13 00:01:33.862738 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 13 00:01:33.862746 systemd[1]: Reached target sockets.target - Socket Units. Oct 13 00:01:33.862754 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Oct 13 00:01:33.862762 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 13 00:01:33.862772 systemd[1]: Finished network-cleanup.service - Network Cleanup. Oct 13 00:01:33.862780 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Oct 13 00:01:33.862788 systemd[1]: Starting systemd-fsck-usr.service... Oct 13 00:01:33.862796 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 13 00:01:33.862804 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 13 00:01:33.862812 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 00:01:33.862820 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Oct 13 00:01:33.862830 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 13 00:01:33.862838 systemd[1]: Finished systemd-fsck-usr.service. Oct 13 00:01:33.862846 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 13 00:01:33.862880 systemd-journald[244]: Collecting audit messages is disabled. Oct 13 00:01:33.862903 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 00:01:33.862911 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Oct 13 00:01:33.862920 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 13 00:01:33.862929 kernel: Bridge firewalling registered Oct 13 00:01:33.862937 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 13 00:01:33.862995 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 13 00:01:33.863010 systemd-journald[244]: Journal started Oct 13 00:01:33.863029 systemd-journald[244]: Runtime Journal (/run/log/journal/e888d1f984e84e75bd58bc92ce2c02ee) is 8M, max 76.5M, 68.5M free. Oct 13 00:01:33.829050 systemd-modules-load[246]: Inserted module 'overlay' Oct 13 00:01:33.855033 systemd-modules-load[246]: Inserted module 'br_netfilter' Oct 13 00:01:33.866165 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 13 00:01:33.873463 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 13 00:01:33.877808 systemd[1]: Started systemd-journald.service - Journal Service. Oct 13 00:01:33.885725 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 13 00:01:33.890248 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 13 00:01:33.893362 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 13 00:01:33.903722 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 13 00:01:33.911668 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Oct 13 00:01:33.916547 systemd-tmpfiles[279]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Oct 13 00:01:33.920159 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 13 00:01:33.928783 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 13 00:01:33.948895 dracut-cmdline[284]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=37fc523060a9b8894388e25ab0f082059dd744d472a2b8577211d4b3dd66a910 Oct 13 00:01:33.978295 systemd-resolved[286]: Positive Trust Anchors: Oct 13 00:01:33.978313 systemd-resolved[286]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 13 00:01:33.978346 systemd-resolved[286]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 13 00:01:33.985341 systemd-resolved[286]: Defaulting to hostname 'linux'. Oct 13 00:01:33.986466 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 13 00:01:33.987177 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 13 00:01:34.072706 kernel: SCSI subsystem initialized Oct 13 00:01:34.077462 kernel: Loading iSCSI transport class v2.0-870. Oct 13 00:01:34.086725 kernel: iscsi: registered transport (tcp) Oct 13 00:01:34.101695 kernel: iscsi: registered transport (qla4xxx) Oct 13 00:01:34.101773 kernel: QLogic iSCSI HBA Driver Oct 13 00:01:34.124914 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 13 00:01:34.157275 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 13 00:01:34.162554 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 13 00:01:34.223459 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Oct 13 00:01:34.226363 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Oct 13 00:01:34.303530 kernel: raid6: neonx8 gen() 15553 MB/s Oct 13 00:01:34.320502 kernel: raid6: neonx4 gen() 15687 MB/s Oct 13 00:01:34.337472 kernel: raid6: neonx2 gen() 13133 MB/s Oct 13 00:01:34.354461 kernel: raid6: neonx1 gen() 10382 MB/s Oct 13 00:01:34.371477 kernel: raid6: int64x8 gen() 6832 MB/s Oct 13 00:01:34.388461 kernel: raid6: int64x4 gen() 7267 MB/s Oct 13 00:01:34.406769 kernel: raid6: int64x2 gen() 6033 MB/s Oct 13 00:01:34.422471 kernel: raid6: int64x1 gen() 4961 MB/s Oct 13 00:01:34.422545 kernel: raid6: using algorithm neonx4 gen() 15687 MB/s Oct 13 00:01:34.439490 kernel: raid6: .... xor() 12270 MB/s, rmw enabled Oct 13 00:01:34.439564 kernel: raid6: using neon recovery algorithm Oct 13 00:01:34.445745 kernel: xor: measuring software checksum speed Oct 13 00:01:34.445824 kernel: 8regs : 21449 MB/sec Oct 13 00:01:34.445838 kernel: 32regs : 20480 MB/sec Oct 13 00:01:34.445851 kernel: arm64_neon : 27589 MB/sec Oct 13 00:01:34.445864 kernel: xor: using function: arm64_neon (27589 MB/sec) Oct 13 00:01:34.505778 kernel: Btrfs loaded, zoned=no, fsverity=no Oct 13 00:01:34.518076 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Oct 13 00:01:34.521789 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 13 00:01:34.551790 systemd-udevd[494]: Using default interface naming scheme 'v255'. Oct 13 00:01:34.558033 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 13 00:01:34.562347 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Oct 13 00:01:34.600697 dracut-pre-trigger[500]: rd.md=0: removing MD RAID activation Oct 13 00:01:34.635393 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Oct 13 00:01:34.637772 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 13 00:01:34.706557 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 13 00:01:34.709161 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Oct 13 00:01:34.824550 kernel: ACPI: bus type USB registered Oct 13 00:01:34.824601 kernel: usbcore: registered new interface driver usbfs Oct 13 00:01:34.824612 kernel: usbcore: registered new interface driver hub Oct 13 00:01:34.828738 kernel: virtio_scsi virtio5: 2/0/0 default/read/poll queues Oct 13 00:01:34.829043 kernel: scsi host0: Virtio SCSI HBA Oct 13 00:01:34.833455 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU CD-ROM 2.5+ PQ: 0 ANSI: 5 Oct 13 00:01:34.833554 kernel: scsi 0:0:0:1: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Oct 13 00:01:34.837439 kernel: usbcore: registered new device driver usb Oct 13 00:01:34.873440 kernel: sr 0:0:0:0: Power-on or device reset occurred Oct 13 00:01:34.875444 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 16x/50x cd/rw xa/form2 cdda tray Oct 13 00:01:34.875659 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Oct 13 00:01:34.877696 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 13 00:01:34.879630 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 00:01:34.881143 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 00:01:34.883106 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Oct 13 00:01:34.883279 kernel: sd 0:0:0:1: Power-on or device reset occurred Oct 13 00:01:34.883370 kernel: sd 0:0:0:1: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Oct 13 00:01:34.885018 kernel: sd 0:0:0:1: [sda] Write Protect is off Oct 13 00:01:34.885189 kernel: sd 0:0:0:1: [sda] Mode Sense: 63 00 00 08 Oct 13 00:01:34.885289 kernel: sd 0:0:0:1: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Oct 13 00:01:34.885920 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 00:01:34.888785 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Oct 13 00:01:34.899646 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Oct 13 00:01:34.899711 kernel: GPT:17805311 != 80003071 Oct 13 00:01:34.899726 kernel: GPT:Alternate GPT header not at the end of the disk. Oct 13 00:01:34.899736 kernel: GPT:17805311 != 80003071 Oct 13 00:01:34.901436 kernel: GPT: Use GNU Parted to correct GPT errors. Oct 13 00:01:34.903096 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Oct 13 00:01:34.903142 kernel: sd 0:0:0:1: [sda] Attached SCSI disk Oct 13 00:01:34.909211 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Oct 13 00:01:34.909441 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Oct 13 00:01:34.909539 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Oct 13 00:01:34.911682 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Oct 13 00:01:34.911886 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Oct 13 00:01:34.913445 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Oct 13 00:01:34.918098 kernel: hub 1-0:1.0: USB hub found Oct 13 00:01:34.918333 kernel: hub 1-0:1.0: 4 ports detected Oct 13 00:01:34.919512 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Oct 13 00:01:34.920756 kernel: hub 2-0:1.0: USB hub found Oct 13 00:01:34.920976 kernel: hub 2-0:1.0: 4 ports detected Oct 13 00:01:34.925747 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 00:01:34.973213 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Oct 13 00:01:34.987803 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Oct 13 00:01:35.027589 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Oct 13 00:01:35.035918 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Oct 13 00:01:35.036746 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Oct 13 00:01:35.044315 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Oct 13 00:01:35.045346 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Oct 13 00:01:35.046860 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Oct 13 00:01:35.047710 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 13 00:01:35.049002 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 13 00:01:35.051177 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Oct 13 00:01:35.069211 disk-uuid[599]: Primary Header is updated. Oct 13 00:01:35.069211 disk-uuid[599]: Secondary Entries is updated. Oct 13 00:01:35.069211 disk-uuid[599]: Secondary Header is updated. Oct 13 00:01:35.079468 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Oct 13 00:01:35.079898 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Oct 13 00:01:35.162439 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Oct 13 00:01:35.296618 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input1 Oct 13 00:01:35.296685 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Oct 13 00:01:35.296883 kernel: usbcore: registered new interface driver usbhid Oct 13 00:01:35.298221 kernel: usbhid: USB HID core driver Oct 13 00:01:35.400478 kernel: usb 1-2: new high-speed USB device number 3 using xhci_hcd Oct 13 00:01:35.527446 kernel: input: QEMU QEMU USB Keyboard as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-2/1-2:1.0/0003:0627:0001.0002/input/input2 Oct 13 00:01:35.579435 kernel: hid-generic 0003:0627:0001.0002: input,hidraw1: USB HID v1.11 Keyboard [QEMU QEMU USB Keyboard] on usb-0000:02:00.0-2/input0 Oct 13 00:01:36.099519 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Oct 13 00:01:36.100535 disk-uuid[601]: The operation has completed successfully. Oct 13 00:01:36.158777 systemd[1]: disk-uuid.service: Deactivated successfully. Oct 13 00:01:36.158895 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Oct 13 00:01:36.188789 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Oct 13 00:01:36.211230 sh[623]: Success Oct 13 00:01:36.227802 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Oct 13 00:01:36.227860 kernel: device-mapper: uevent: version 1.0.3 Oct 13 00:01:36.227871 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Oct 13 00:01:36.235472 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Oct 13 00:01:36.284042 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Oct 13 00:01:36.289454 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Oct 13 00:01:36.301460 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Oct 13 00:01:36.317966 kernel: BTRFS: device fsid e4495086-3456-43e0-be7b-4c3c53a67174 devid 1 transid 38 /dev/mapper/usr (254:0) scanned by mount (635) Oct 13 00:01:36.321087 kernel: BTRFS info (device dm-0): first mount of filesystem e4495086-3456-43e0-be7b-4c3c53a67174 Oct 13 00:01:36.321147 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Oct 13 00:01:36.328460 kernel: BTRFS info (device dm-0): enabling ssd optimizations Oct 13 00:01:36.328526 kernel: BTRFS info (device dm-0): disabling log replay at mount time Oct 13 00:01:36.329423 kernel: BTRFS info (device dm-0): enabling free space tree Oct 13 00:01:36.330987 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Oct 13 00:01:36.332607 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Oct 13 00:01:36.334402 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Oct 13 00:01:36.336655 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Oct 13 00:01:36.339150 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Oct 13 00:01:36.380484 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (664) Oct 13 00:01:36.382084 kernel: BTRFS info (device sda6): first mount of filesystem 51f6bef3-5c80-492f-be85-d924f50fa726 Oct 13 00:01:36.382247 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Oct 13 00:01:36.387673 kernel: BTRFS info (device sda6): enabling ssd optimizations Oct 13 00:01:36.387746 kernel: BTRFS info (device sda6): turning on async discard Oct 13 00:01:36.388453 kernel: BTRFS info (device sda6): enabling free space tree Oct 13 00:01:36.394468 kernel: BTRFS info (device sda6): last unmount of filesystem 51f6bef3-5c80-492f-be85-d924f50fa726 Oct 13 00:01:36.396154 systemd[1]: Finished ignition-setup.service - Ignition (setup). Oct 13 00:01:36.399886 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Oct 13 00:01:36.489978 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 13 00:01:36.492398 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 13 00:01:36.534682 systemd-networkd[806]: lo: Link UP Oct 13 00:01:36.534698 systemd-networkd[806]: lo: Gained carrier Oct 13 00:01:36.536777 systemd-networkd[806]: Enumeration completed Oct 13 00:01:36.536898 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 13 00:01:36.538295 systemd-networkd[806]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:36.538299 systemd-networkd[806]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 13 00:01:36.539419 systemd-networkd[806]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:36.539598 systemd[1]: Reached target network.target - Network. Oct 13 00:01:36.540591 systemd-networkd[806]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 13 00:01:36.540987 systemd-networkd[806]: eth0: Link UP Oct 13 00:01:36.541191 systemd-networkd[806]: eth1: Link UP Oct 13 00:01:36.541345 systemd-networkd[806]: eth0: Gained carrier Oct 13 00:01:36.541355 systemd-networkd[806]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:36.546213 systemd-networkd[806]: eth1: Gained carrier Oct 13 00:01:36.546231 systemd-networkd[806]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:36.571636 ignition[715]: Ignition 2.22.0 Oct 13 00:01:36.571656 ignition[715]: Stage: fetch-offline Oct 13 00:01:36.571719 ignition[715]: no configs at "/usr/lib/ignition/base.d" Oct 13 00:01:36.571730 ignition[715]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Oct 13 00:01:36.571831 ignition[715]: parsed url from cmdline: "" Oct 13 00:01:36.575755 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Oct 13 00:01:36.571834 ignition[715]: no config URL provided Oct 13 00:01:36.577653 systemd-networkd[806]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Oct 13 00:01:36.571850 ignition[715]: reading system config file "/usr/lib/ignition/user.ign" Oct 13 00:01:36.579029 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Oct 13 00:01:36.571857 ignition[715]: no config at "/usr/lib/ignition/user.ign" Oct 13 00:01:36.571865 ignition[715]: failed to fetch config: resource requires networking Oct 13 00:01:36.572061 ignition[715]: Ignition finished successfully Oct 13 00:01:36.591552 systemd-networkd[806]: eth0: DHCPv4 address 49.13.69.12/32, gateway 172.31.1.1 acquired from 172.31.1.1 Oct 13 00:01:36.612554 ignition[816]: Ignition 2.22.0 Oct 13 00:01:36.612566 ignition[816]: Stage: fetch Oct 13 00:01:36.612701 ignition[816]: no configs at "/usr/lib/ignition/base.d" Oct 13 00:01:36.612718 ignition[816]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Oct 13 00:01:36.612800 ignition[816]: parsed url from cmdline: "" Oct 13 00:01:36.612804 ignition[816]: no config URL provided Oct 13 00:01:36.612809 ignition[816]: reading system config file "/usr/lib/ignition/user.ign" Oct 13 00:01:36.612818 ignition[816]: no config at "/usr/lib/ignition/user.ign" Oct 13 00:01:36.612844 ignition[816]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Oct 13 00:01:36.619998 ignition[816]: GET result: OK Oct 13 00:01:36.620544 ignition[816]: parsing config with SHA512: d05748392b11c29a8741e1317e658cf6a0223aad01ad194710d3a6d2afeb500ea09e047ac4776e8d884d9a812c59bc5e6fcb9e946f2d592e95630445ea5cbbed Oct 13 00:01:36.627179 unknown[816]: fetched base config from "system" Oct 13 00:01:36.627200 unknown[816]: fetched base config from "system" Oct 13 00:01:36.627210 unknown[816]: fetched user config from "hetzner" Oct 13 00:01:36.630554 ignition[816]: fetch: fetch complete Oct 13 00:01:36.630564 ignition[816]: fetch: fetch passed Oct 13 00:01:36.631160 ignition[816]: Ignition finished successfully Oct 13 00:01:36.634543 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Oct 13 00:01:36.636906 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Oct 13 00:01:36.679839 ignition[822]: Ignition 2.22.0 Oct 13 00:01:36.680505 ignition[822]: Stage: kargs Oct 13 00:01:36.680681 ignition[822]: no configs at "/usr/lib/ignition/base.d" Oct 13 00:01:36.680690 ignition[822]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Oct 13 00:01:36.683213 ignition[822]: kargs: kargs passed Oct 13 00:01:36.683277 ignition[822]: Ignition finished successfully Oct 13 00:01:36.686367 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Oct 13 00:01:36.689081 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Oct 13 00:01:36.739056 ignition[829]: Ignition 2.22.0 Oct 13 00:01:36.739075 ignition[829]: Stage: disks Oct 13 00:01:36.739219 ignition[829]: no configs at "/usr/lib/ignition/base.d" Oct 13 00:01:36.739228 ignition[829]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Oct 13 00:01:36.740197 ignition[829]: disks: disks passed Oct 13 00:01:36.740251 ignition[829]: Ignition finished successfully Oct 13 00:01:36.745358 systemd[1]: Finished ignition-disks.service - Ignition (disks). Oct 13 00:01:36.746591 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Oct 13 00:01:36.747324 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 13 00:01:36.748238 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 13 00:01:36.749391 systemd[1]: Reached target sysinit.target - System Initialization. Oct 13 00:01:36.752611 systemd[1]: Reached target basic.target - Basic System. Oct 13 00:01:36.755279 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Oct 13 00:01:36.803700 systemd-fsck[838]: ROOT: clean, 15/1628000 files, 120826/1617920 blocks Oct 13 00:01:36.807131 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Oct 13 00:01:36.811599 systemd[1]: Mounting sysroot.mount - /sysroot... Oct 13 00:01:36.899467 kernel: EXT4-fs (sda9): mounted filesystem 1aa1d0b4-cbac-4728-b9e0-662fa574e9ad r/w with ordered data mode. Quota mode: none. Oct 13 00:01:36.901004 systemd[1]: Mounted sysroot.mount - /sysroot. Oct 13 00:01:36.903115 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Oct 13 00:01:36.906131 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 13 00:01:36.909530 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Oct 13 00:01:36.917692 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Oct 13 00:01:36.920533 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Oct 13 00:01:36.922549 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Oct 13 00:01:36.926114 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Oct 13 00:01:36.928444 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (846) Oct 13 00:01:36.929998 kernel: BTRFS info (device sda6): first mount of filesystem 51f6bef3-5c80-492f-be85-d924f50fa726 Oct 13 00:01:36.930038 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Oct 13 00:01:36.931647 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Oct 13 00:01:36.938698 kernel: BTRFS info (device sda6): enabling ssd optimizations Oct 13 00:01:36.938761 kernel: BTRFS info (device sda6): turning on async discard Oct 13 00:01:36.939451 kernel: BTRFS info (device sda6): enabling free space tree Oct 13 00:01:36.948224 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 13 00:01:36.982569 initrd-setup-root[874]: cut: /sysroot/etc/passwd: No such file or directory Oct 13 00:01:36.990172 coreos-metadata[848]: Oct 13 00:01:36.990 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Oct 13 00:01:36.994357 coreos-metadata[848]: Oct 13 00:01:36.994 INFO Fetch successful Oct 13 00:01:36.995661 coreos-metadata[848]: Oct 13 00:01:36.995 INFO wrote hostname ci-4459-1-0-3-bea61e6062 to /sysroot/etc/hostname Oct 13 00:01:36.997109 initrd-setup-root[881]: cut: /sysroot/etc/group: No such file or directory Oct 13 00:01:37.001429 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Oct 13 00:01:37.005157 initrd-setup-root[889]: cut: /sysroot/etc/shadow: No such file or directory Oct 13 00:01:37.010300 initrd-setup-root[896]: cut: /sysroot/etc/gshadow: No such file or directory Oct 13 00:01:37.115520 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Oct 13 00:01:37.119553 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Oct 13 00:01:37.121823 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Oct 13 00:01:37.136459 kernel: BTRFS info (device sda6): last unmount of filesystem 51f6bef3-5c80-492f-be85-d924f50fa726 Oct 13 00:01:37.158648 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Oct 13 00:01:37.168095 ignition[964]: INFO : Ignition 2.22.0 Oct 13 00:01:37.168868 ignition[964]: INFO : Stage: mount Oct 13 00:01:37.169493 ignition[964]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 13 00:01:37.171003 ignition[964]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Oct 13 00:01:37.171003 ignition[964]: INFO : mount: mount passed Oct 13 00:01:37.171003 ignition[964]: INFO : Ignition finished successfully Oct 13 00:01:37.173612 systemd[1]: Finished ignition-mount.service - Ignition (mount). Oct 13 00:01:37.177504 systemd[1]: Starting ignition-files.service - Ignition (files)... Oct 13 00:01:37.317234 systemd[1]: sysroot-oem.mount: Deactivated successfully. Oct 13 00:01:37.320945 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 13 00:01:37.349494 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (975) Oct 13 00:01:37.350708 kernel: BTRFS info (device sda6): first mount of filesystem 51f6bef3-5c80-492f-be85-d924f50fa726 Oct 13 00:01:37.350739 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Oct 13 00:01:37.355232 kernel: BTRFS info (device sda6): enabling ssd optimizations Oct 13 00:01:37.355303 kernel: BTRFS info (device sda6): turning on async discard Oct 13 00:01:37.355314 kernel: BTRFS info (device sda6): enabling free space tree Oct 13 00:01:37.358348 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 13 00:01:37.404472 ignition[992]: INFO : Ignition 2.22.0 Oct 13 00:01:37.406689 ignition[992]: INFO : Stage: files Oct 13 00:01:37.406689 ignition[992]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 13 00:01:37.406689 ignition[992]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Oct 13 00:01:37.406689 ignition[992]: DEBUG : files: compiled without relabeling support, skipping Oct 13 00:01:37.409243 ignition[992]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Oct 13 00:01:37.409243 ignition[992]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Oct 13 00:01:37.411779 ignition[992]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Oct 13 00:01:37.412984 ignition[992]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Oct 13 00:01:37.414395 unknown[992]: wrote ssh authorized keys file for user: core Oct 13 00:01:37.415680 ignition[992]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Oct 13 00:01:37.417795 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Oct 13 00:01:37.420186 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 Oct 13 00:01:37.524462 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Oct 13 00:01:37.617089 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 13 00:01:37.618727 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 13 00:01:37.628815 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Oct 13 00:01:37.628815 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Oct 13 00:01:37.628815 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Oct 13 00:01:37.628815 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Oct 13 00:01:37.628815 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Oct 13 00:01:37.628815 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-arm64.raw: attempt #1 Oct 13 00:01:37.854714 systemd-networkd[806]: eth0: Gained IPv6LL Oct 13 00:01:38.018880 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Oct 13 00:01:38.174665 systemd-networkd[806]: eth1: Gained IPv6LL Oct 13 00:01:39.149893 ignition[992]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Oct 13 00:01:39.149893 ignition[992]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Oct 13 00:01:39.153354 ignition[992]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 13 00:01:39.156797 ignition[992]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 13 00:01:39.156797 ignition[992]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Oct 13 00:01:39.156797 ignition[992]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Oct 13 00:01:39.160421 ignition[992]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Oct 13 00:01:39.160421 ignition[992]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Oct 13 00:01:39.160421 ignition[992]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Oct 13 00:01:39.160421 ignition[992]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Oct 13 00:01:39.160421 ignition[992]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Oct 13 00:01:39.160421 ignition[992]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Oct 13 00:01:39.160421 ignition[992]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Oct 13 00:01:39.160421 ignition[992]: INFO : files: files passed Oct 13 00:01:39.160421 ignition[992]: INFO : Ignition finished successfully Oct 13 00:01:39.160002 systemd[1]: Finished ignition-files.service - Ignition (files). Oct 13 00:01:39.163558 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Oct 13 00:01:39.167020 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Oct 13 00:01:39.185571 systemd[1]: ignition-quench.service: Deactivated successfully. Oct 13 00:01:39.186779 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Oct 13 00:01:39.193652 initrd-setup-root-after-ignition[1022]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 13 00:01:39.193652 initrd-setup-root-after-ignition[1022]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Oct 13 00:01:39.195635 initrd-setup-root-after-ignition[1026]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 13 00:01:39.199447 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 13 00:01:39.200691 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Oct 13 00:01:39.203585 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Oct 13 00:01:39.266314 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Oct 13 00:01:39.266491 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Oct 13 00:01:39.268333 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Oct 13 00:01:39.269578 systemd[1]: Reached target initrd.target - Initrd Default Target. Oct 13 00:01:39.271005 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Oct 13 00:01:39.272114 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Oct 13 00:01:39.302690 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 13 00:01:39.306109 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Oct 13 00:01:39.329783 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Oct 13 00:01:39.330649 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 13 00:01:39.332679 systemd[1]: Stopped target timers.target - Timer Units. Oct 13 00:01:39.334517 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Oct 13 00:01:39.334867 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 13 00:01:39.337691 systemd[1]: Stopped target initrd.target - Initrd Default Target. Oct 13 00:01:39.338362 systemd[1]: Stopped target basic.target - Basic System. Oct 13 00:01:39.339374 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Oct 13 00:01:39.340469 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Oct 13 00:01:39.341741 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Oct 13 00:01:39.343012 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Oct 13 00:01:39.344316 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Oct 13 00:01:39.345564 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Oct 13 00:01:39.346804 systemd[1]: Stopped target sysinit.target - System Initialization. Oct 13 00:01:39.348094 systemd[1]: Stopped target local-fs.target - Local File Systems. Oct 13 00:01:39.349166 systemd[1]: Stopped target swap.target - Swaps. Oct 13 00:01:39.350194 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Oct 13 00:01:39.350325 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Oct 13 00:01:39.351575 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Oct 13 00:01:39.352271 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 13 00:01:39.353379 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Oct 13 00:01:39.356499 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 13 00:01:39.358217 systemd[1]: dracut-initqueue.service: Deactivated successfully. Oct 13 00:01:39.358386 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Oct 13 00:01:39.361075 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Oct 13 00:01:39.361233 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 13 00:01:39.363240 systemd[1]: ignition-files.service: Deactivated successfully. Oct 13 00:01:39.363348 systemd[1]: Stopped ignition-files.service - Ignition (files). Oct 13 00:01:39.364702 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Oct 13 00:01:39.364809 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Oct 13 00:01:39.366938 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Oct 13 00:01:39.369042 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Oct 13 00:01:39.369210 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Oct 13 00:01:39.374722 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Oct 13 00:01:39.375211 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Oct 13 00:01:39.375341 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Oct 13 00:01:39.379300 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Oct 13 00:01:39.379439 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Oct 13 00:01:39.389723 systemd[1]: initrd-cleanup.service: Deactivated successfully. Oct 13 00:01:39.389837 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Oct 13 00:01:39.406911 systemd[1]: sysroot-boot.mount: Deactivated successfully. Oct 13 00:01:39.415554 systemd[1]: sysroot-boot.service: Deactivated successfully. Oct 13 00:01:39.416190 ignition[1046]: INFO : Ignition 2.22.0 Oct 13 00:01:39.416190 ignition[1046]: INFO : Stage: umount Oct 13 00:01:39.417300 ignition[1046]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 13 00:01:39.417300 ignition[1046]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Oct 13 00:01:39.417300 ignition[1046]: INFO : umount: umount passed Oct 13 00:01:39.417300 ignition[1046]: INFO : Ignition finished successfully Oct 13 00:01:39.417244 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Oct 13 00:01:39.419956 systemd[1]: ignition-mount.service: Deactivated successfully. Oct 13 00:01:39.420089 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Oct 13 00:01:39.421039 systemd[1]: ignition-disks.service: Deactivated successfully. Oct 13 00:01:39.421107 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Oct 13 00:01:39.421850 systemd[1]: ignition-kargs.service: Deactivated successfully. Oct 13 00:01:39.421933 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Oct 13 00:01:39.423554 systemd[1]: ignition-fetch.service: Deactivated successfully. Oct 13 00:01:39.423654 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Oct 13 00:01:39.424979 systemd[1]: Stopped target network.target - Network. Oct 13 00:01:39.425997 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Oct 13 00:01:39.426070 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Oct 13 00:01:39.427166 systemd[1]: Stopped target paths.target - Path Units. Oct 13 00:01:39.428051 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Oct 13 00:01:39.431484 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 13 00:01:39.432632 systemd[1]: Stopped target slices.target - Slice Units. Oct 13 00:01:39.434109 systemd[1]: Stopped target sockets.target - Socket Units. Oct 13 00:01:39.435381 systemd[1]: iscsid.socket: Deactivated successfully. Oct 13 00:01:39.435463 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Oct 13 00:01:39.436471 systemd[1]: iscsiuio.socket: Deactivated successfully. Oct 13 00:01:39.436507 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 13 00:01:39.437779 systemd[1]: ignition-setup.service: Deactivated successfully. Oct 13 00:01:39.437876 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Oct 13 00:01:39.438808 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Oct 13 00:01:39.438845 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Oct 13 00:01:39.439775 systemd[1]: initrd-setup-root.service: Deactivated successfully. Oct 13 00:01:39.439834 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Oct 13 00:01:39.440880 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Oct 13 00:01:39.441891 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Oct 13 00:01:39.447609 systemd[1]: systemd-resolved.service: Deactivated successfully. Oct 13 00:01:39.450014 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Oct 13 00:01:39.455810 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Oct 13 00:01:39.457387 systemd[1]: systemd-networkd.service: Deactivated successfully. Oct 13 00:01:39.457589 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Oct 13 00:01:39.460257 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Oct 13 00:01:39.460513 systemd[1]: Stopped target network-pre.target - Preparation for Network. Oct 13 00:01:39.462179 systemd[1]: systemd-networkd.socket: Deactivated successfully. Oct 13 00:01:39.462221 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Oct 13 00:01:39.464304 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Oct 13 00:01:39.466151 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Oct 13 00:01:39.466243 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 13 00:01:39.469556 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 13 00:01:39.469626 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 13 00:01:39.471974 systemd[1]: systemd-modules-load.service: Deactivated successfully. Oct 13 00:01:39.472044 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Oct 13 00:01:39.473307 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Oct 13 00:01:39.473356 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 13 00:01:39.475081 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 13 00:01:39.481483 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Oct 13 00:01:39.481582 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Oct 13 00:01:39.494788 systemd[1]: systemd-udevd.service: Deactivated successfully. Oct 13 00:01:39.496183 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 13 00:01:39.497444 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Oct 13 00:01:39.497498 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Oct 13 00:01:39.498353 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Oct 13 00:01:39.498403 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Oct 13 00:01:39.501356 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Oct 13 00:01:39.501517 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Oct 13 00:01:39.504743 systemd[1]: dracut-cmdline.service: Deactivated successfully. Oct 13 00:01:39.504813 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Oct 13 00:01:39.507030 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 13 00:01:39.507101 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 13 00:01:39.510735 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Oct 13 00:01:39.512227 systemd[1]: systemd-network-generator.service: Deactivated successfully. Oct 13 00:01:39.512306 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Oct 13 00:01:39.516060 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Oct 13 00:01:39.516125 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 13 00:01:39.519117 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 13 00:01:39.519180 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 00:01:39.523668 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Oct 13 00:01:39.523764 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Oct 13 00:01:39.523810 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Oct 13 00:01:39.524178 systemd[1]: network-cleanup.service: Deactivated successfully. Oct 13 00:01:39.526514 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Oct 13 00:01:39.532959 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Oct 13 00:01:39.533065 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Oct 13 00:01:39.535913 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Oct 13 00:01:39.537309 systemd[1]: Starting initrd-switch-root.service - Switch Root... Oct 13 00:01:39.559788 systemd[1]: Switching root. Oct 13 00:01:39.596471 systemd-journald[244]: Journal stopped Oct 13 00:01:40.673488 systemd-journald[244]: Received SIGTERM from PID 1 (systemd). Oct 13 00:01:40.673556 kernel: SELinux: policy capability network_peer_controls=1 Oct 13 00:01:40.673568 kernel: SELinux: policy capability open_perms=1 Oct 13 00:01:40.673577 kernel: SELinux: policy capability extended_socket_class=1 Oct 13 00:01:40.673586 kernel: SELinux: policy capability always_check_network=0 Oct 13 00:01:40.673595 kernel: SELinux: policy capability cgroup_seclabel=1 Oct 13 00:01:40.673610 kernel: SELinux: policy capability nnp_nosuid_transition=1 Oct 13 00:01:40.673620 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Oct 13 00:01:40.673628 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Oct 13 00:01:40.673637 kernel: SELinux: policy capability userspace_initial_context=0 Oct 13 00:01:40.673646 kernel: audit: type=1403 audit(1760313699.761:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Oct 13 00:01:40.673659 systemd[1]: Successfully loaded SELinux policy in 71.836ms. Oct 13 00:01:40.673683 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 7.127ms. Oct 13 00:01:40.673694 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 13 00:01:40.673707 systemd[1]: Detected virtualization kvm. Oct 13 00:01:40.673716 systemd[1]: Detected architecture arm64. Oct 13 00:01:40.673726 systemd[1]: Detected first boot. Oct 13 00:01:40.673736 systemd[1]: Hostname set to . Oct 13 00:01:40.673745 systemd[1]: Initializing machine ID from VM UUID. Oct 13 00:01:40.673754 zram_generator::config[1090]: No configuration found. Oct 13 00:01:40.673765 kernel: NET: Registered PF_VSOCK protocol family Oct 13 00:01:40.673777 systemd[1]: Populated /etc with preset unit settings. Oct 13 00:01:40.673791 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Oct 13 00:01:40.673803 systemd[1]: initrd-switch-root.service: Deactivated successfully. Oct 13 00:01:40.673813 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Oct 13 00:01:40.673823 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Oct 13 00:01:40.673834 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Oct 13 00:01:40.673845 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Oct 13 00:01:40.673856 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Oct 13 00:01:40.673865 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Oct 13 00:01:40.673875 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Oct 13 00:01:40.673899 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Oct 13 00:01:40.673912 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Oct 13 00:01:40.673922 systemd[1]: Created slice user.slice - User and Session Slice. Oct 13 00:01:40.673933 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 13 00:01:40.673943 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 13 00:01:40.673953 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Oct 13 00:01:40.673965 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Oct 13 00:01:40.673976 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Oct 13 00:01:40.673986 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 13 00:01:40.673996 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Oct 13 00:01:40.674006 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 13 00:01:40.674019 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 13 00:01:40.674030 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Oct 13 00:01:40.674040 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Oct 13 00:01:40.674051 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Oct 13 00:01:40.674061 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Oct 13 00:01:40.674071 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 13 00:01:40.674084 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 13 00:01:40.674095 systemd[1]: Reached target slices.target - Slice Units. Oct 13 00:01:40.674104 systemd[1]: Reached target swap.target - Swaps. Oct 13 00:01:40.674114 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Oct 13 00:01:40.674126 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Oct 13 00:01:40.674136 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Oct 13 00:01:40.674146 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 13 00:01:40.674156 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 13 00:01:40.674171 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 13 00:01:40.674183 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Oct 13 00:01:40.674197 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Oct 13 00:01:40.674210 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Oct 13 00:01:40.674221 systemd[1]: Mounting media.mount - External Media Directory... Oct 13 00:01:40.674239 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Oct 13 00:01:40.674251 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Oct 13 00:01:40.674264 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Oct 13 00:01:40.674276 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Oct 13 00:01:40.674288 systemd[1]: Reached target machines.target - Containers. Oct 13 00:01:40.674297 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Oct 13 00:01:40.674307 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 13 00:01:40.674318 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 13 00:01:40.674330 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Oct 13 00:01:40.674340 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 13 00:01:40.674349 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 13 00:01:40.674359 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 13 00:01:40.674370 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Oct 13 00:01:40.674379 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 13 00:01:40.674389 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 13 00:01:40.674400 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Oct 13 00:01:40.676356 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Oct 13 00:01:40.676377 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Oct 13 00:01:40.676387 systemd[1]: Stopped systemd-fsck-usr.service. Oct 13 00:01:40.676398 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 13 00:01:40.676422 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 13 00:01:40.676437 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 13 00:01:40.676447 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 13 00:01:40.677484 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Oct 13 00:01:40.677503 kernel: fuse: init (API version 7.41) Oct 13 00:01:40.677514 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Oct 13 00:01:40.677526 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 13 00:01:40.677542 systemd[1]: verity-setup.service: Deactivated successfully. Oct 13 00:01:40.677552 systemd[1]: Stopped verity-setup.service. Oct 13 00:01:40.677562 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Oct 13 00:01:40.677574 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Oct 13 00:01:40.677584 systemd[1]: Mounted media.mount - External Media Directory. Oct 13 00:01:40.677595 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Oct 13 00:01:40.677605 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Oct 13 00:01:40.677615 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Oct 13 00:01:40.677626 kernel: ACPI: bus type drm_connector registered Oct 13 00:01:40.677637 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 13 00:01:40.677648 systemd[1]: modprobe@configfs.service: Deactivated successfully. Oct 13 00:01:40.677658 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Oct 13 00:01:40.677668 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 13 00:01:40.677677 kernel: loop: module loaded Oct 13 00:01:40.677687 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 13 00:01:40.677698 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 13 00:01:40.677708 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 13 00:01:40.677720 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 13 00:01:40.677730 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 13 00:01:40.677740 systemd[1]: modprobe@fuse.service: Deactivated successfully. Oct 13 00:01:40.677751 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Oct 13 00:01:40.677761 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Oct 13 00:01:40.677771 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 13 00:01:40.677781 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 13 00:01:40.677791 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 13 00:01:40.677808 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 13 00:01:40.677822 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Oct 13 00:01:40.677834 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Oct 13 00:01:40.677846 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 13 00:01:40.677858 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Oct 13 00:01:40.677985 systemd-journald[1154]: Collecting audit messages is disabled. Oct 13 00:01:40.678024 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Oct 13 00:01:40.678039 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 13 00:01:40.678051 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 13 00:01:40.678065 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Oct 13 00:01:40.678076 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Oct 13 00:01:40.678093 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 13 00:01:40.678106 systemd-journald[1154]: Journal started Oct 13 00:01:40.678127 systemd-journald[1154]: Runtime Journal (/run/log/journal/e888d1f984e84e75bd58bc92ce2c02ee) is 8M, max 76.5M, 68.5M free. Oct 13 00:01:40.321807 systemd[1]: Queued start job for default target multi-user.target. Oct 13 00:01:40.349399 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Oct 13 00:01:40.349987 systemd[1]: systemd-journald.service: Deactivated successfully. Oct 13 00:01:40.683517 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Oct 13 00:01:40.683564 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 13 00:01:40.686443 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Oct 13 00:01:40.686519 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 13 00:01:40.695490 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 13 00:01:40.698465 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Oct 13 00:01:40.703855 systemd[1]: Starting systemd-sysusers.service - Create System Users... Oct 13 00:01:40.710520 systemd[1]: Started systemd-journald.service - Journal Service. Oct 13 00:01:40.708687 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Oct 13 00:01:40.709813 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Oct 13 00:01:40.748566 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Oct 13 00:01:40.751655 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Oct 13 00:01:40.753264 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Oct 13 00:01:40.757545 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Oct 13 00:01:40.764585 kernel: loop0: detected capacity change from 0 to 119368 Oct 13 00:01:40.779065 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 13 00:01:40.799228 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Oct 13 00:01:40.801547 systemd-journald[1154]: Time spent on flushing to /var/log/journal/e888d1f984e84e75bd58bc92ce2c02ee is 85.483ms for 1179 entries. Oct 13 00:01:40.801547 systemd-journald[1154]: System Journal (/var/log/journal/e888d1f984e84e75bd58bc92ce2c02ee) is 8M, max 584.8M, 576.8M free. Oct 13 00:01:40.901846 systemd-journald[1154]: Received client request to flush runtime journal. Oct 13 00:01:40.901970 kernel: loop1: detected capacity change from 0 to 207008 Oct 13 00:01:40.902005 kernel: loop2: detected capacity change from 0 to 100632 Oct 13 00:01:40.821923 systemd[1]: Finished systemd-sysusers.service - Create System Users. Oct 13 00:01:40.826185 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 13 00:01:40.831836 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 13 00:01:40.875533 systemd-tmpfiles[1223]: ACLs are not supported, ignoring. Oct 13 00:01:40.875546 systemd-tmpfiles[1223]: ACLs are not supported, ignoring. Oct 13 00:01:40.886583 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 13 00:01:40.907481 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Oct 13 00:01:40.909660 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Oct 13 00:01:40.925504 kernel: loop3: detected capacity change from 0 to 8 Oct 13 00:01:40.940474 kernel: loop4: detected capacity change from 0 to 119368 Oct 13 00:01:40.976427 kernel: loop5: detected capacity change from 0 to 207008 Oct 13 00:01:40.999478 kernel: loop6: detected capacity change from 0 to 100632 Oct 13 00:01:41.025436 kernel: loop7: detected capacity change from 0 to 8 Oct 13 00:01:41.026765 (sd-merge)[1234]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Oct 13 00:01:41.027990 (sd-merge)[1234]: Merged extensions into '/usr'. Oct 13 00:01:41.035558 systemd[1]: Reload requested from client PID 1191 ('systemd-sysext') (unit systemd-sysext.service)... Oct 13 00:01:41.035584 systemd[1]: Reloading... Oct 13 00:01:41.174551 zram_generator::config[1257]: No configuration found. Oct 13 00:01:41.271705 ldconfig[1187]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Oct 13 00:01:41.395775 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Oct 13 00:01:41.396504 systemd[1]: Reloading finished in 360 ms. Oct 13 00:01:41.417798 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Oct 13 00:01:41.420483 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Oct 13 00:01:41.431630 systemd[1]: Starting ensure-sysext.service... Oct 13 00:01:41.439783 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 13 00:01:41.462223 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Oct 13 00:01:41.467589 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 13 00:01:41.474548 systemd[1]: Reload requested from client PID 1298 ('systemctl') (unit ensure-sysext.service)... Oct 13 00:01:41.474568 systemd[1]: Reloading... Oct 13 00:01:41.474832 systemd-tmpfiles[1299]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Oct 13 00:01:41.474853 systemd-tmpfiles[1299]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Oct 13 00:01:41.475116 systemd-tmpfiles[1299]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Oct 13 00:01:41.475297 systemd-tmpfiles[1299]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Oct 13 00:01:41.476075 systemd-tmpfiles[1299]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Oct 13 00:01:41.476277 systemd-tmpfiles[1299]: ACLs are not supported, ignoring. Oct 13 00:01:41.476319 systemd-tmpfiles[1299]: ACLs are not supported, ignoring. Oct 13 00:01:41.480461 systemd-tmpfiles[1299]: Detected autofs mount point /boot during canonicalization of boot. Oct 13 00:01:41.480471 systemd-tmpfiles[1299]: Skipping /boot Oct 13 00:01:41.488729 systemd-tmpfiles[1299]: Detected autofs mount point /boot during canonicalization of boot. Oct 13 00:01:41.488743 systemd-tmpfiles[1299]: Skipping /boot Oct 13 00:01:41.536146 systemd-udevd[1302]: Using default interface naming scheme 'v255'. Oct 13 00:01:41.566456 zram_generator::config[1327]: No configuration found. Oct 13 00:01:41.856612 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Oct 13 00:01:41.857020 systemd[1]: Reloading finished in 382 ms. Oct 13 00:01:41.866001 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 13 00:01:41.867853 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 13 00:01:41.928524 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 13 00:01:41.932731 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Oct 13 00:01:41.934702 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 13 00:01:41.938788 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 13 00:01:41.941756 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 13 00:01:41.958451 kernel: mousedev: PS/2 mouse device common for all mice Oct 13 00:01:41.964686 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 13 00:01:41.966625 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 13 00:01:41.966763 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 13 00:01:41.968265 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Oct 13 00:01:41.974991 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 13 00:01:41.981782 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 13 00:01:41.988946 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Oct 13 00:01:41.991168 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 13 00:01:41.991743 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 13 00:01:41.998085 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 13 00:01:41.998321 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 13 00:01:42.022211 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Oct 13 00:01:42.032114 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 13 00:01:42.032771 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 13 00:01:42.035015 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 13 00:01:42.037463 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 13 00:01:42.039820 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 13 00:01:42.047798 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 13 00:01:42.048650 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 13 00:01:42.048776 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 13 00:01:42.051344 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Oct 13 00:01:42.062468 systemd[1]: Finished ensure-sysext.service. Oct 13 00:01:42.068719 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Oct 13 00:01:42.070953 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Oct 13 00:01:42.072857 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Oct 13 00:01:42.078674 systemd[1]: Starting systemd-update-done.service - Update is Completed... Oct 13 00:01:42.109259 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 13 00:01:42.109495 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 13 00:01:42.111136 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 13 00:01:42.114863 systemd[1]: Finished systemd-update-done.service - Update is Completed. Oct 13 00:01:42.117869 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 13 00:01:42.118211 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 13 00:01:42.125251 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 13 00:01:42.125490 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 13 00:01:42.126423 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 13 00:01:42.139949 augenrules[1457]: No rules Oct 13 00:01:42.142308 systemd[1]: audit-rules.service: Deactivated successfully. Oct 13 00:01:42.143176 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 13 00:01:42.163359 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Oct 13 00:01:42.164703 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 13 00:01:42.277846 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Oct 13 00:01:42.279727 systemd[1]: Started systemd-userdbd.service - User Database Manager. Oct 13 00:01:42.294121 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Oct 13 00:01:42.298892 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 00:01:42.311689 kernel: [drm] pci: virtio-gpu-pci detected at 0000:00:01.0 Oct 13 00:01:42.311806 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Oct 13 00:01:42.311826 kernel: [drm] features: -context_init Oct 13 00:01:42.316447 kernel: [drm] number of scanouts: 1 Oct 13 00:01:42.316539 kernel: [drm] number of cap sets: 0 Oct 13 00:01:42.323438 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:01.0 on minor 0 Oct 13 00:01:42.332551 kernel: Console: switching to colour frame buffer device 160x50 Oct 13 00:01:42.345467 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Oct 13 00:01:42.354052 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Oct 13 00:01:42.367660 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 13 00:01:42.367930 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 00:01:42.374617 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 00:01:42.442780 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Oct 13 00:01:42.443978 systemd[1]: Reached target time-set.target - System Time Set. Oct 13 00:01:42.450191 systemd-networkd[1418]: lo: Link UP Oct 13 00:01:42.450198 systemd-networkd[1418]: lo: Gained carrier Oct 13 00:01:42.453858 systemd-networkd[1418]: Enumeration completed Oct 13 00:01:42.454015 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 13 00:01:42.454651 systemd-networkd[1418]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:42.454656 systemd-networkd[1418]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 13 00:01:42.456007 systemd-networkd[1418]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:42.456019 systemd-networkd[1418]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 13 00:01:42.457160 systemd-networkd[1418]: eth0: Link UP Oct 13 00:01:42.457472 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Oct 13 00:01:42.458822 systemd-resolved[1420]: Positive Trust Anchors: Oct 13 00:01:42.458834 systemd-networkd[1418]: eth0: Gained carrier Oct 13 00:01:42.458859 systemd-networkd[1418]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:42.459866 systemd-resolved[1420]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 13 00:01:42.459930 systemd-resolved[1420]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 13 00:01:42.460702 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Oct 13 00:01:42.463758 systemd-networkd[1418]: eth1: Link UP Oct 13 00:01:42.464627 systemd-networkd[1418]: eth1: Gained carrier Oct 13 00:01:42.464652 systemd-networkd[1418]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 00:01:42.468521 systemd-resolved[1420]: Using system hostname 'ci-4459-1-0-3-bea61e6062'. Oct 13 00:01:42.471728 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 13 00:01:42.472524 systemd[1]: Reached target network.target - Network. Oct 13 00:01:42.473026 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 13 00:01:42.487153 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 00:01:42.488204 systemd[1]: Reached target sysinit.target - System Initialization. Oct 13 00:01:42.490954 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Oct 13 00:01:42.491862 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Oct 13 00:01:42.493071 systemd[1]: Started logrotate.timer - Daily rotation of log files. Oct 13 00:01:42.494024 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Oct 13 00:01:42.494996 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Oct 13 00:01:42.495568 systemd-networkd[1418]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Oct 13 00:01:42.495923 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Oct 13 00:01:42.495954 systemd[1]: Reached target paths.target - Path Units. Oct 13 00:01:42.496969 systemd[1]: Reached target timers.target - Timer Units. Oct 13 00:01:42.497741 systemd-timesyncd[1441]: Network configuration changed, trying to establish connection. Oct 13 00:01:42.499800 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Oct 13 00:01:42.502284 systemd[1]: Starting docker.socket - Docker Socket for the API... Oct 13 00:01:42.505625 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Oct 13 00:01:42.506623 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Oct 13 00:01:42.507334 systemd[1]: Reached target ssh-access.target - SSH Access Available. Oct 13 00:01:42.518597 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Oct 13 00:01:42.520545 systemd-networkd[1418]: eth0: DHCPv4 address 49.13.69.12/32, gateway 172.31.1.1 acquired from 172.31.1.1 Oct 13 00:01:42.521336 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Oct 13 00:01:42.521630 systemd-timesyncd[1441]: Network configuration changed, trying to establish connection. Oct 13 00:01:42.524278 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Oct 13 00:01:42.525530 systemd[1]: Listening on docker.socket - Docker Socket for the API. Oct 13 00:01:42.529102 systemd[1]: Reached target sockets.target - Socket Units. Oct 13 00:01:42.530044 systemd[1]: Reached target basic.target - Basic System. Oct 13 00:01:42.530660 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Oct 13 00:01:42.530696 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Oct 13 00:01:42.531931 systemd[1]: Starting containerd.service - containerd container runtime... Oct 13 00:01:42.536650 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Oct 13 00:01:42.542971 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Oct 13 00:01:42.546023 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Oct 13 00:01:42.554717 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Oct 13 00:01:42.559256 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Oct 13 00:01:42.559977 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Oct 13 00:01:42.562296 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Oct 13 00:01:42.572549 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Oct 13 00:01:42.577652 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Oct 13 00:01:42.582734 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Oct 13 00:01:42.587691 jq[1509]: false Oct 13 00:01:42.589088 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Oct 13 00:01:42.595764 systemd[1]: Starting systemd-logind.service - User Login Management... Oct 13 00:01:42.598491 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Oct 13 00:01:42.599048 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Oct 13 00:01:42.600297 systemd[1]: Starting update-engine.service - Update Engine... Oct 13 00:01:42.610031 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Oct 13 00:01:42.610403 coreos-metadata[1506]: Oct 13 00:01:42.610 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Oct 13 00:01:42.612864 coreos-metadata[1506]: Oct 13 00:01:42.612 INFO Fetch successful Oct 13 00:01:42.615436 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Oct 13 00:01:42.618571 coreos-metadata[1506]: Oct 13 00:01:42.615 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Oct 13 00:01:42.618571 coreos-metadata[1506]: Oct 13 00:01:42.616 INFO Fetch successful Oct 13 00:01:42.616534 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Oct 13 00:01:42.620504 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Oct 13 00:01:42.631998 extend-filesystems[1510]: Found /dev/sda6 Oct 13 00:01:42.654836 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Oct 13 00:01:42.655118 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Oct 13 00:01:42.656474 systemd[1]: motdgen.service: Deactivated successfully. Oct 13 00:01:42.658504 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Oct 13 00:01:42.662641 tar[1525]: linux-arm64/LICENSE Oct 13 00:01:42.663129 tar[1525]: linux-arm64/helm Oct 13 00:01:42.671020 extend-filesystems[1510]: Found /dev/sda9 Oct 13 00:01:42.676837 jq[1523]: true Oct 13 00:01:42.683132 extend-filesystems[1510]: Checking size of /dev/sda9 Oct 13 00:01:42.689670 update_engine[1522]: I20251013 00:01:42.688362 1522 main.cc:92] Flatcar Update Engine starting Oct 13 00:01:42.699819 (ntainerd)[1544]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Oct 13 00:01:42.716987 extend-filesystems[1510]: Resized partition /dev/sda9 Oct 13 00:01:42.727947 extend-filesystems[1557]: resize2fs 1.47.3 (8-Jul-2025) Oct 13 00:01:42.729599 jq[1549]: true Oct 13 00:01:42.740937 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Oct 13 00:01:42.748606 dbus-daemon[1507]: [system] SELinux support is enabled Oct 13 00:01:42.749374 systemd[1]: Started dbus.service - D-Bus System Message Bus. Oct 13 00:01:42.753586 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Oct 13 00:01:42.753628 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Oct 13 00:01:42.755523 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Oct 13 00:01:42.755564 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Oct 13 00:01:42.788258 systemd[1]: Started update-engine.service - Update Engine. Oct 13 00:01:42.789554 update_engine[1522]: I20251013 00:01:42.789362 1522 update_check_scheduler.cc:74] Next update check in 2m55s Oct 13 00:01:42.802997 systemd[1]: Started locksmithd.service - Cluster reboot manager. Oct 13 00:01:42.855207 systemd-logind[1521]: New seat seat0. Oct 13 00:01:42.892575 systemd-logind[1521]: Watching system buttons on /dev/input/event0 (Power Button) Oct 13 00:01:42.892602 systemd-logind[1521]: Watching system buttons on /dev/input/event2 (QEMU QEMU USB Keyboard) Oct 13 00:01:42.893657 systemd[1]: Started systemd-logind.service - User Login Management. Oct 13 00:01:42.936605 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Oct 13 00:01:42.939501 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Oct 13 00:01:42.968004 bash[1581]: Updated "/home/core/.ssh/authorized_keys" Oct 13 00:01:42.971332 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Oct 13 00:01:42.974486 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Oct 13 00:01:42.974225 systemd[1]: Starting sshkeys.service... Oct 13 00:01:42.996298 extend-filesystems[1557]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Oct 13 00:01:42.996298 extend-filesystems[1557]: old_desc_blocks = 1, new_desc_blocks = 5 Oct 13 00:01:42.996298 extend-filesystems[1557]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Oct 13 00:01:43.004776 extend-filesystems[1510]: Resized filesystem in /dev/sda9 Oct 13 00:01:42.998838 systemd[1]: extend-filesystems.service: Deactivated successfully. Oct 13 00:01:43.000545 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Oct 13 00:01:43.026680 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Oct 13 00:01:43.030012 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Oct 13 00:01:43.088331 coreos-metadata[1591]: Oct 13 00:01:43.088 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Oct 13 00:01:43.089922 coreos-metadata[1591]: Oct 13 00:01:43.089 INFO Fetch successful Oct 13 00:01:43.097147 unknown[1591]: wrote ssh authorized keys file for user: core Oct 13 00:01:43.149199 locksmithd[1564]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Oct 13 00:01:43.157686 containerd[1544]: time="2025-10-13T00:01:43Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Oct 13 00:01:43.160568 containerd[1544]: time="2025-10-13T00:01:43.160510680Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Oct 13 00:01:43.161022 update-ssh-keys[1600]: Updated "/home/core/.ssh/authorized_keys" Oct 13 00:01:43.163835 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Oct 13 00:01:43.168890 systemd[1]: Finished sshkeys.service. Oct 13 00:01:43.179233 containerd[1544]: time="2025-10-13T00:01:43.178895080Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.84µs" Oct 13 00:01:43.179233 containerd[1544]: time="2025-10-13T00:01:43.178946960Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Oct 13 00:01:43.179233 containerd[1544]: time="2025-10-13T00:01:43.178969120Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Oct 13 00:01:43.179233 containerd[1544]: time="2025-10-13T00:01:43.179150560Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Oct 13 00:01:43.179233 containerd[1544]: time="2025-10-13T00:01:43.179167160Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Oct 13 00:01:43.179233 containerd[1544]: time="2025-10-13T00:01:43.179194960Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 13 00:01:43.179731 containerd[1544]: time="2025-10-13T00:01:43.179254680Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 13 00:01:43.179731 containerd[1544]: time="2025-10-13T00:01:43.179271000Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 13 00:01:43.179731 containerd[1544]: time="2025-10-13T00:01:43.179572640Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 13 00:01:43.179731 containerd[1544]: time="2025-10-13T00:01:43.179589440Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 13 00:01:43.179731 containerd[1544]: time="2025-10-13T00:01:43.179611040Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 13 00:01:43.179731 containerd[1544]: time="2025-10-13T00:01:43.179620960Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Oct 13 00:01:43.179731 containerd[1544]: time="2025-10-13T00:01:43.179709320Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Oct 13 00:01:43.180735 containerd[1544]: time="2025-10-13T00:01:43.179957800Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 13 00:01:43.180735 containerd[1544]: time="2025-10-13T00:01:43.180000240Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 13 00:01:43.180735 containerd[1544]: time="2025-10-13T00:01:43.180013440Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Oct 13 00:01:43.180735 containerd[1544]: time="2025-10-13T00:01:43.180063200Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Oct 13 00:01:43.180735 containerd[1544]: time="2025-10-13T00:01:43.180448760Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Oct 13 00:01:43.180735 containerd[1544]: time="2025-10-13T00:01:43.180518240Z" level=info msg="metadata content store policy set" policy=shared Oct 13 00:01:43.188837 containerd[1544]: time="2025-10-13T00:01:43.188776720Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Oct 13 00:01:43.188975 containerd[1544]: time="2025-10-13T00:01:43.188910240Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Oct 13 00:01:43.189003 containerd[1544]: time="2025-10-13T00:01:43.188983960Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Oct 13 00:01:43.189022 containerd[1544]: time="2025-10-13T00:01:43.189000240Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Oct 13 00:01:43.189022 containerd[1544]: time="2025-10-13T00:01:43.189014640Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Oct 13 00:01:43.189073 containerd[1544]: time="2025-10-13T00:01:43.189028320Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Oct 13 00:01:43.189073 containerd[1544]: time="2025-10-13T00:01:43.189042080Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Oct 13 00:01:43.189125 containerd[1544]: time="2025-10-13T00:01:43.189056680Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Oct 13 00:01:43.189125 containerd[1544]: time="2025-10-13T00:01:43.189092200Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Oct 13 00:01:43.189125 containerd[1544]: time="2025-10-13T00:01:43.189103760Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Oct 13 00:01:43.189125 containerd[1544]: time="2025-10-13T00:01:43.189114080Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Oct 13 00:01:43.189195 containerd[1544]: time="2025-10-13T00:01:43.189128400Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Oct 13 00:01:43.189356 containerd[1544]: time="2025-10-13T00:01:43.189300840Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Oct 13 00:01:43.189356 containerd[1544]: time="2025-10-13T00:01:43.189330880Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Oct 13 00:01:43.189356 containerd[1544]: time="2025-10-13T00:01:43.189348960Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Oct 13 00:01:43.189597 containerd[1544]: time="2025-10-13T00:01:43.189394800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Oct 13 00:01:43.190453 containerd[1544]: time="2025-10-13T00:01:43.190429640Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Oct 13 00:01:43.190504 containerd[1544]: time="2025-10-13T00:01:43.190459960Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Oct 13 00:01:43.190504 containerd[1544]: time="2025-10-13T00:01:43.190475800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Oct 13 00:01:43.190504 containerd[1544]: time="2025-10-13T00:01:43.190487440Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Oct 13 00:01:43.190504 containerd[1544]: time="2025-10-13T00:01:43.190500600Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Oct 13 00:01:43.190711 containerd[1544]: time="2025-10-13T00:01:43.190511640Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Oct 13 00:01:43.190711 containerd[1544]: time="2025-10-13T00:01:43.190524440Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Oct 13 00:01:43.190747 containerd[1544]: time="2025-10-13T00:01:43.190717720Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Oct 13 00:01:43.190747 containerd[1544]: time="2025-10-13T00:01:43.190732400Z" level=info msg="Start snapshots syncer" Oct 13 00:01:43.191059 containerd[1544]: time="2025-10-13T00:01:43.190766480Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Oct 13 00:01:43.191597 containerd[1544]: time="2025-10-13T00:01:43.191554000Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Oct 13 00:01:43.191860 containerd[1544]: time="2025-10-13T00:01:43.191624160Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Oct 13 00:01:43.192159 containerd[1544]: time="2025-10-13T00:01:43.192132360Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Oct 13 00:01:43.192319 containerd[1544]: time="2025-10-13T00:01:43.192299120Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Oct 13 00:01:43.192359 containerd[1544]: time="2025-10-13T00:01:43.192334720Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Oct 13 00:01:43.192359 containerd[1544]: time="2025-10-13T00:01:43.192350120Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Oct 13 00:01:43.192397 containerd[1544]: time="2025-10-13T00:01:43.192365160Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Oct 13 00:01:43.192397 containerd[1544]: time="2025-10-13T00:01:43.192381880Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.192401840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197481640Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197550200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197565920Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197580360Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197620800Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197637520Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197647120Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197656960Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197664880Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197673600Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197684600Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197766160Z" level=info msg="runtime interface created" Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197771320Z" level=info msg="created NRI interface" Oct 13 00:01:43.199587 containerd[1544]: time="2025-10-13T00:01:43.197781800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Oct 13 00:01:43.199962 containerd[1544]: time="2025-10-13T00:01:43.197798600Z" level=info msg="Connect containerd service" Oct 13 00:01:43.199962 containerd[1544]: time="2025-10-13T00:01:43.197831040Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Oct 13 00:01:43.201997 containerd[1544]: time="2025-10-13T00:01:43.201954280Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 13 00:01:43.390163 containerd[1544]: time="2025-10-13T00:01:43.390023120Z" level=info msg="Start subscribing containerd event" Oct 13 00:01:43.390163 containerd[1544]: time="2025-10-13T00:01:43.390100080Z" level=info msg="Start recovering state" Oct 13 00:01:43.390280 containerd[1544]: time="2025-10-13T00:01:43.390200800Z" level=info msg="Start event monitor" Oct 13 00:01:43.390280 containerd[1544]: time="2025-10-13T00:01:43.390214200Z" level=info msg="Start cni network conf syncer for default" Oct 13 00:01:43.390280 containerd[1544]: time="2025-10-13T00:01:43.390222480Z" level=info msg="Start streaming server" Oct 13 00:01:43.390280 containerd[1544]: time="2025-10-13T00:01:43.390231200Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Oct 13 00:01:43.390280 containerd[1544]: time="2025-10-13T00:01:43.390237600Z" level=info msg="runtime interface starting up..." Oct 13 00:01:43.390280 containerd[1544]: time="2025-10-13T00:01:43.390242840Z" level=info msg="starting plugins..." Oct 13 00:01:43.390280 containerd[1544]: time="2025-10-13T00:01:43.390256440Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Oct 13 00:01:43.392793 containerd[1544]: time="2025-10-13T00:01:43.392093200Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Oct 13 00:01:43.392793 containerd[1544]: time="2025-10-13T00:01:43.392171680Z" level=info msg=serving... address=/run/containerd/containerd.sock Oct 13 00:01:43.392793 containerd[1544]: time="2025-10-13T00:01:43.392238760Z" level=info msg="containerd successfully booted in 0.235085s" Oct 13 00:01:43.392360 systemd[1]: Started containerd.service - containerd container runtime. Oct 13 00:01:43.437708 tar[1525]: linux-arm64/README.md Oct 13 00:01:43.458624 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Oct 13 00:01:43.550572 systemd-networkd[1418]: eth0: Gained IPv6LL Oct 13 00:01:43.551182 systemd-timesyncd[1441]: Network configuration changed, trying to establish connection. Oct 13 00:01:43.554841 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Oct 13 00:01:43.558232 systemd[1]: Reached target network-online.target - Network is Online. Oct 13 00:01:43.562630 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:01:43.566823 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Oct 13 00:01:43.629290 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Oct 13 00:01:43.934574 systemd-networkd[1418]: eth1: Gained IPv6LL Oct 13 00:01:43.935460 systemd-timesyncd[1441]: Network configuration changed, trying to establish connection. Oct 13 00:01:44.386608 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:01:44.397127 (kubelet)[1637]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 00:01:44.883937 kubelet[1637]: E1013 00:01:44.883779 1637 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 00:01:44.887835 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 00:01:44.888063 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 00:01:44.890526 systemd[1]: kubelet.service: Consumed 857ms CPU time, 254.9M memory peak. Oct 13 00:01:45.988587 sshd_keygen[1552]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Oct 13 00:01:46.014976 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Oct 13 00:01:46.021684 systemd[1]: Starting issuegen.service - Generate /run/issue... Oct 13 00:01:46.045470 systemd[1]: issuegen.service: Deactivated successfully. Oct 13 00:01:46.045743 systemd[1]: Finished issuegen.service - Generate /run/issue. Oct 13 00:01:46.048390 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Oct 13 00:01:46.075372 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Oct 13 00:01:46.080587 systemd[1]: Started getty@tty1.service - Getty on tty1. Oct 13 00:01:46.085016 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Oct 13 00:01:46.087539 systemd[1]: Reached target getty.target - Login Prompts. Oct 13 00:01:46.088579 systemd[1]: Reached target multi-user.target - Multi-User System. Oct 13 00:01:46.092555 systemd[1]: Startup finished in 2.399s (kernel) + 6.126s (initrd) + 6.400s (userspace) = 14.926s. Oct 13 00:01:55.138990 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Oct 13 00:01:55.141116 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:01:55.344814 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:01:55.356393 (kubelet)[1673]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 00:01:55.411529 kubelet[1673]: E1013 00:01:55.411371 1673 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 00:01:55.415612 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 00:01:55.415832 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 00:01:55.416601 systemd[1]: kubelet.service: Consumed 201ms CPU time, 105M memory peak. Oct 13 00:02:05.666402 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Oct 13 00:02:05.669585 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:05.872498 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:05.885375 (kubelet)[1688]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 00:02:05.937661 kubelet[1688]: E1013 00:02:05.937492 1688 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 00:02:05.940648 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 00:02:05.940840 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 00:02:05.941621 systemd[1]: kubelet.service: Consumed 222ms CPU time, 107.5M memory peak. Oct 13 00:02:13.897977 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Oct 13 00:02:13.900729 systemd[1]: Started sshd@0-49.13.69.12:22-139.178.89.65:47454.service - OpenSSH per-connection server daemon (139.178.89.65:47454). Oct 13 00:02:14.203757 systemd-timesyncd[1441]: Contacted time server 217.144.138.234:123 (2.flatcar.pool.ntp.org). Oct 13 00:02:14.203918 systemd-timesyncd[1441]: Initial clock synchronization to Mon 2025-10-13 00:02:14.401026 UTC. Oct 13 00:02:14.896106 sshd[1696]: Accepted publickey for core from 139.178.89.65 port 47454 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:02:14.902028 sshd-session[1696]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:02:14.913631 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Oct 13 00:02:14.915376 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Oct 13 00:02:14.931462 systemd-logind[1521]: New session 1 of user core. Oct 13 00:02:14.947294 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Oct 13 00:02:14.952221 systemd[1]: Starting user@500.service - User Manager for UID 500... Oct 13 00:02:14.969393 (systemd)[1701]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Oct 13 00:02:14.974200 systemd-logind[1521]: New session c1 of user core. Oct 13 00:02:15.136674 systemd[1701]: Queued start job for default target default.target. Oct 13 00:02:15.144605 systemd[1701]: Created slice app.slice - User Application Slice. Oct 13 00:02:15.144664 systemd[1701]: Reached target paths.target - Paths. Oct 13 00:02:15.144731 systemd[1701]: Reached target timers.target - Timers. Oct 13 00:02:15.147085 systemd[1701]: Starting dbus.socket - D-Bus User Message Bus Socket... Oct 13 00:02:15.165695 systemd[1701]: Listening on dbus.socket - D-Bus User Message Bus Socket. Oct 13 00:02:15.166157 systemd[1701]: Reached target sockets.target - Sockets. Oct 13 00:02:15.166349 systemd[1701]: Reached target basic.target - Basic System. Oct 13 00:02:15.166543 systemd[1701]: Reached target default.target - Main User Target. Oct 13 00:02:15.166590 systemd[1]: Started user@500.service - User Manager for UID 500. Oct 13 00:02:15.166776 systemd[1701]: Startup finished in 183ms. Oct 13 00:02:15.174230 systemd[1]: Started session-1.scope - Session 1 of User core. Oct 13 00:02:15.868783 systemd[1]: Started sshd@1-49.13.69.12:22-139.178.89.65:47458.service - OpenSSH per-connection server daemon (139.178.89.65:47458). Oct 13 00:02:16.191648 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Oct 13 00:02:16.196318 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:16.402510 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:16.417831 (kubelet)[1723]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 00:02:16.472776 kubelet[1723]: E1013 00:02:16.472607 1723 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 00:02:16.476901 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 00:02:16.477063 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 00:02:16.479500 systemd[1]: kubelet.service: Consumed 192ms CPU time, 105.5M memory peak. Oct 13 00:02:16.898542 sshd[1712]: Accepted publickey for core from 139.178.89.65 port 47458 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:02:16.902366 sshd-session[1712]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:02:16.912679 systemd-logind[1521]: New session 2 of user core. Oct 13 00:02:16.921367 systemd[1]: Started session-2.scope - Session 2 of User core. Oct 13 00:02:17.586156 sshd[1730]: Connection closed by 139.178.89.65 port 47458 Oct 13 00:02:17.587892 sshd-session[1712]: pam_unix(sshd:session): session closed for user core Oct 13 00:02:17.594231 systemd-logind[1521]: Session 2 logged out. Waiting for processes to exit. Oct 13 00:02:17.596530 systemd[1]: sshd@1-49.13.69.12:22-139.178.89.65:47458.service: Deactivated successfully. Oct 13 00:02:17.602014 systemd[1]: session-2.scope: Deactivated successfully. Oct 13 00:02:17.607754 systemd-logind[1521]: Removed session 2. Oct 13 00:02:17.758843 systemd[1]: Started sshd@2-49.13.69.12:22-139.178.89.65:47460.service - OpenSSH per-connection server daemon (139.178.89.65:47460). Oct 13 00:02:18.769320 sshd[1736]: Accepted publickey for core from 139.178.89.65 port 47460 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:02:18.771905 sshd-session[1736]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:02:18.781564 systemd-logind[1521]: New session 3 of user core. Oct 13 00:02:18.788869 systemd[1]: Started session-3.scope - Session 3 of User core. Oct 13 00:02:19.451541 sshd[1739]: Connection closed by 139.178.89.65 port 47460 Oct 13 00:02:19.452261 sshd-session[1736]: pam_unix(sshd:session): session closed for user core Oct 13 00:02:19.463216 systemd-logind[1521]: Session 3 logged out. Waiting for processes to exit. Oct 13 00:02:19.464197 systemd[1]: sshd@2-49.13.69.12:22-139.178.89.65:47460.service: Deactivated successfully. Oct 13 00:02:19.470002 systemd[1]: session-3.scope: Deactivated successfully. Oct 13 00:02:19.473409 systemd-logind[1521]: Removed session 3. Oct 13 00:02:19.629324 systemd[1]: Started sshd@3-49.13.69.12:22-139.178.89.65:47468.service - OpenSSH per-connection server daemon (139.178.89.65:47468). Oct 13 00:02:20.651540 sshd[1745]: Accepted publickey for core from 139.178.89.65 port 47468 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:02:20.655840 sshd-session[1745]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:02:20.662022 systemd-logind[1521]: New session 4 of user core. Oct 13 00:02:20.670323 systemd[1]: Started session-4.scope - Session 4 of User core. Oct 13 00:02:21.336456 sshd[1748]: Connection closed by 139.178.89.65 port 47468 Oct 13 00:02:21.337374 sshd-session[1745]: pam_unix(sshd:session): session closed for user core Oct 13 00:02:21.344047 systemd[1]: sshd@3-49.13.69.12:22-139.178.89.65:47468.service: Deactivated successfully. Oct 13 00:02:21.346460 systemd[1]: session-4.scope: Deactivated successfully. Oct 13 00:02:21.352245 systemd-logind[1521]: Session 4 logged out. Waiting for processes to exit. Oct 13 00:02:21.356139 systemd-logind[1521]: Removed session 4. Oct 13 00:02:21.507295 systemd[1]: Started sshd@4-49.13.69.12:22-139.178.89.65:47472.service - OpenSSH per-connection server daemon (139.178.89.65:47472). Oct 13 00:02:22.501485 sshd[1754]: Accepted publickey for core from 139.178.89.65 port 47472 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:02:22.505341 sshd-session[1754]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:02:22.514816 systemd-logind[1521]: New session 5 of user core. Oct 13 00:02:22.518680 systemd[1]: Started session-5.scope - Session 5 of User core. Oct 13 00:02:23.024206 sudo[1758]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Oct 13 00:02:23.024543 sudo[1758]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 00:02:23.045051 sudo[1758]: pam_unix(sudo:session): session closed for user root Oct 13 00:02:23.202164 sshd[1757]: Connection closed by 139.178.89.65 port 47472 Oct 13 00:02:23.205717 sshd-session[1754]: pam_unix(sshd:session): session closed for user core Oct 13 00:02:23.215573 systemd[1]: sshd@4-49.13.69.12:22-139.178.89.65:47472.service: Deactivated successfully. Oct 13 00:02:23.220612 systemd[1]: session-5.scope: Deactivated successfully. Oct 13 00:02:23.227437 systemd-logind[1521]: Session 5 logged out. Waiting for processes to exit. Oct 13 00:02:23.229677 systemd-logind[1521]: Removed session 5. Oct 13 00:02:23.376275 systemd[1]: Started sshd@5-49.13.69.12:22-139.178.89.65:47266.service - OpenSSH per-connection server daemon (139.178.89.65:47266). Oct 13 00:02:24.363087 sshd[1764]: Accepted publickey for core from 139.178.89.65 port 47266 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:02:24.367394 sshd-session[1764]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:02:24.375078 systemd-logind[1521]: New session 6 of user core. Oct 13 00:02:24.384001 systemd[1]: Started session-6.scope - Session 6 of User core. Oct 13 00:02:24.885336 sudo[1769]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Oct 13 00:02:24.886745 sudo[1769]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 00:02:24.898205 sudo[1769]: pam_unix(sudo:session): session closed for user root Oct 13 00:02:24.908364 sudo[1768]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Oct 13 00:02:24.909718 sudo[1768]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 00:02:24.927240 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 13 00:02:24.985338 augenrules[1791]: No rules Oct 13 00:02:24.986689 systemd[1]: audit-rules.service: Deactivated successfully. Oct 13 00:02:24.987181 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 13 00:02:24.990157 sudo[1768]: pam_unix(sudo:session): session closed for user root Oct 13 00:02:25.150294 sshd[1767]: Connection closed by 139.178.89.65 port 47266 Oct 13 00:02:25.150936 sshd-session[1764]: pam_unix(sshd:session): session closed for user core Oct 13 00:02:25.155605 systemd-logind[1521]: Session 6 logged out. Waiting for processes to exit. Oct 13 00:02:25.157207 systemd[1]: sshd@5-49.13.69.12:22-139.178.89.65:47266.service: Deactivated successfully. Oct 13 00:02:25.160547 systemd[1]: session-6.scope: Deactivated successfully. Oct 13 00:02:25.166042 systemd-logind[1521]: Removed session 6. Oct 13 00:02:25.314800 systemd[1]: Started sshd@6-49.13.69.12:22-139.178.89.65:47278.service - OpenSSH per-connection server daemon (139.178.89.65:47278). Oct 13 00:02:26.322977 sshd[1800]: Accepted publickey for core from 139.178.89.65 port 47278 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:02:26.325339 sshd-session[1800]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:02:26.333300 systemd-logind[1521]: New session 7 of user core. Oct 13 00:02:26.338799 systemd[1]: Started session-7.scope - Session 7 of User core. Oct 13 00:02:26.591918 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Oct 13 00:02:26.600002 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:26.772666 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:26.784037 (kubelet)[1812]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 00:02:26.832083 kubelet[1812]: E1013 00:02:26.832002 1812 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 00:02:26.835788 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 00:02:26.836117 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 00:02:26.836928 systemd[1]: kubelet.service: Consumed 182ms CPU time, 107M memory peak. Oct 13 00:02:26.839159 sudo[1818]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Oct 13 00:02:26.839584 sudo[1818]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 00:02:27.211762 systemd[1]: Starting docker.service - Docker Application Container Engine... Oct 13 00:02:27.239590 (dockerd)[1837]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Oct 13 00:02:27.500539 dockerd[1837]: time="2025-10-13T00:02:27.498698393Z" level=info msg="Starting up" Oct 13 00:02:27.503516 dockerd[1837]: time="2025-10-13T00:02:27.503465413Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Oct 13 00:02:27.519884 dockerd[1837]: time="2025-10-13T00:02:27.519821053Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Oct 13 00:02:27.538494 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport4993426-merged.mount: Deactivated successfully. Oct 13 00:02:27.564272 dockerd[1837]: time="2025-10-13T00:02:27.563582888Z" level=info msg="Loading containers: start." Oct 13 00:02:27.575760 kernel: Initializing XFRM netlink socket Oct 13 00:02:27.853673 systemd-networkd[1418]: docker0: Link UP Oct 13 00:02:27.858952 dockerd[1837]: time="2025-10-13T00:02:27.858824334Z" level=info msg="Loading containers: done." Oct 13 00:02:27.877851 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3732966458-merged.mount: Deactivated successfully. Oct 13 00:02:27.885213 dockerd[1837]: time="2025-10-13T00:02:27.885159783Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Oct 13 00:02:27.885975 dockerd[1837]: time="2025-10-13T00:02:27.885887951Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Oct 13 00:02:27.887225 dockerd[1837]: time="2025-10-13T00:02:27.887184656Z" level=info msg="Initializing buildkit" Oct 13 00:02:27.924254 dockerd[1837]: time="2025-10-13T00:02:27.924176265Z" level=info msg="Completed buildkit initialization" Oct 13 00:02:27.933423 dockerd[1837]: time="2025-10-13T00:02:27.933342563Z" level=info msg="Daemon has completed initialization" Oct 13 00:02:27.933563 dockerd[1837]: time="2025-10-13T00:02:27.933439041Z" level=info msg="API listen on /run/docker.sock" Oct 13 00:02:27.936513 systemd[1]: Started docker.service - Docker Application Container Engine. Oct 13 00:02:28.167772 update_engine[1522]: I20251013 00:02:28.167066 1522 update_attempter.cc:509] Updating boot flags... Oct 13 00:02:29.196871 containerd[1544]: time="2025-10-13T00:02:29.196815702Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Oct 13 00:02:29.898912 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1197744320.mount: Deactivated successfully. Oct 13 00:02:31.005250 containerd[1544]: time="2025-10-13T00:02:31.005174148Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:31.006698 containerd[1544]: time="2025-10-13T00:02:31.006633515Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=26363783" Oct 13 00:02:31.007323 containerd[1544]: time="2025-10-13T00:02:31.007264712Z" level=info msg="ImageCreate event name:\"sha256:02ea53851f07db91ed471dab1ab11541f5c294802371cd8f0cfd423cd5c71002\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:31.011867 containerd[1544]: time="2025-10-13T00:02:31.011783023Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:31.013269 containerd[1544]: time="2025-10-13T00:02:31.013072897Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:02ea53851f07db91ed471dab1ab11541f5c294802371cd8f0cfd423cd5c71002\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"26360284\" in 1.816212635s" Oct 13 00:02:31.013269 containerd[1544]: time="2025-10-13T00:02:31.013138889Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:02ea53851f07db91ed471dab1ab11541f5c294802371cd8f0cfd423cd5c71002\"" Oct 13 00:02:31.014295 containerd[1544]: time="2025-10-13T00:02:31.014218872Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Oct 13 00:02:32.277141 containerd[1544]: time="2025-10-13T00:02:32.275617092Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:32.277141 containerd[1544]: time="2025-10-13T00:02:32.276930278Z" level=info msg="ImageCreate event name:\"sha256:f0bcbad5082c944520b370596a2384affda710b9d7daf84e8a48352699af8e4b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:32.277141 containerd[1544]: time="2025-10-13T00:02:32.276993840Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=22531220" Oct 13 00:02:32.280883 containerd[1544]: time="2025-10-13T00:02:32.280831136Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:32.282013 containerd[1544]: time="2025-10-13T00:02:32.281965465Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:f0bcbad5082c944520b370596a2384affda710b9d7daf84e8a48352699af8e4b\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"24099975\" in 1.2677064s" Oct 13 00:02:32.282013 containerd[1544]: time="2025-10-13T00:02:32.282006490Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:f0bcbad5082c944520b370596a2384affda710b9d7daf84e8a48352699af8e4b\"" Oct 13 00:02:32.282663 containerd[1544]: time="2025-10-13T00:02:32.282626269Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Oct 13 00:02:33.472526 containerd[1544]: time="2025-10-13T00:02:33.472460290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:33.475038 containerd[1544]: time="2025-10-13T00:02:33.474972131Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=17484344" Oct 13 00:02:33.478434 containerd[1544]: time="2025-10-13T00:02:33.478325248Z" level=info msg="ImageCreate event name:\"sha256:1d625baf81b59592006d97a6741bc947698ed222b612ac10efa57b7aa96d2a27\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:33.483121 containerd[1544]: time="2025-10-13T00:02:33.483040322Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:33.484702 containerd[1544]: time="2025-10-13T00:02:33.484165391Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:1d625baf81b59592006d97a6741bc947698ed222b612ac10efa57b7aa96d2a27\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"19053117\" in 1.201499063s" Oct 13 00:02:33.484702 containerd[1544]: time="2025-10-13T00:02:33.484213578Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:1d625baf81b59592006d97a6741bc947698ed222b612ac10efa57b7aa96d2a27\"" Oct 13 00:02:33.485682 containerd[1544]: time="2025-10-13T00:02:33.485635108Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Oct 13 00:02:34.484091 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1735391979.mount: Deactivated successfully. Oct 13 00:02:34.872683 containerd[1544]: time="2025-10-13T00:02:34.872627696Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:34.875225 containerd[1544]: time="2025-10-13T00:02:34.875165730Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=27417843" Oct 13 00:02:34.877234 containerd[1544]: time="2025-10-13T00:02:34.877143990Z" level=info msg="ImageCreate event name:\"sha256:72b57ec14d31e8422925ef4c3eff44822cdc04a11fd30d13824f1897d83a16d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:34.882339 containerd[1544]: time="2025-10-13T00:02:34.881324630Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:34.882791 containerd[1544]: time="2025-10-13T00:02:34.882735543Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:72b57ec14d31e8422925ef4c3eff44822cdc04a11fd30d13824f1897d83a16d4\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"27416836\" in 1.397048723s" Oct 13 00:02:34.882791 containerd[1544]: time="2025-10-13T00:02:34.882777826Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:72b57ec14d31e8422925ef4c3eff44822cdc04a11fd30d13824f1897d83a16d4\"" Oct 13 00:02:34.883834 containerd[1544]: time="2025-10-13T00:02:34.883727679Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Oct 13 00:02:35.505848 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3096666060.mount: Deactivated successfully. Oct 13 00:02:36.213443 containerd[1544]: time="2025-10-13T00:02:36.212600716Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:36.215316 containerd[1544]: time="2025-10-13T00:02:36.215237257Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951714" Oct 13 00:02:36.217215 containerd[1544]: time="2025-10-13T00:02:36.217162815Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:36.221203 containerd[1544]: time="2025-10-13T00:02:36.221112320Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:36.222831 containerd[1544]: time="2025-10-13T00:02:36.222791029Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.33732929s" Oct 13 00:02:36.223117 containerd[1544]: time="2025-10-13T00:02:36.223002625Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Oct 13 00:02:36.225421 containerd[1544]: time="2025-10-13T00:02:36.225323174Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Oct 13 00:02:36.762794 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount858993241.mount: Deactivated successfully. Oct 13 00:02:36.773395 containerd[1544]: time="2025-10-13T00:02:36.771753875Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 13 00:02:36.773395 containerd[1544]: time="2025-10-13T00:02:36.772958316Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268723" Oct 13 00:02:36.773395 containerd[1544]: time="2025-10-13T00:02:36.773290212Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 13 00:02:36.776010 containerd[1544]: time="2025-10-13T00:02:36.775954275Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 13 00:02:36.777141 containerd[1544]: time="2025-10-13T00:02:36.777089812Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 551.660199ms" Oct 13 00:02:36.777141 containerd[1544]: time="2025-10-13T00:02:36.777130994Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Oct 13 00:02:36.777684 containerd[1544]: time="2025-10-13T00:02:36.777652333Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Oct 13 00:02:36.841655 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Oct 13 00:02:36.844810 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:37.025869 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:37.039283 (kubelet)[2202]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 00:02:37.087386 kubelet[2202]: E1013 00:02:37.087309 2202 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 00:02:37.089829 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 00:02:37.090172 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 00:02:37.090796 systemd[1]: kubelet.service: Consumed 184ms CPU time, 105.1M memory peak. Oct 13 00:02:37.356127 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1846982952.mount: Deactivated successfully. Oct 13 00:02:38.972117 containerd[1544]: time="2025-10-13T00:02:38.972064135Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:38.975281 containerd[1544]: time="2025-10-13T00:02:38.975235686Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67943239" Oct 13 00:02:38.977087 containerd[1544]: time="2025-10-13T00:02:38.977027818Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:38.983114 containerd[1544]: time="2025-10-13T00:02:38.982376421Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:02:38.984313 containerd[1544]: time="2025-10-13T00:02:38.984250447Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 2.206556016s" Oct 13 00:02:38.984313 containerd[1544]: time="2025-10-13T00:02:38.984307432Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" Oct 13 00:02:44.906395 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:44.907175 systemd[1]: kubelet.service: Consumed 184ms CPU time, 105.1M memory peak. Oct 13 00:02:44.910659 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:44.947081 systemd[1]: Reload requested from client PID 2291 ('systemctl') (unit session-7.scope)... Oct 13 00:02:44.947106 systemd[1]: Reloading... Oct 13 00:02:45.099438 zram_generator::config[2347]: No configuration found. Oct 13 00:02:45.285826 systemd[1]: Reloading finished in 338 ms. Oct 13 00:02:45.344908 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Oct 13 00:02:45.344995 systemd[1]: kubelet.service: Failed with result 'signal'. Oct 13 00:02:45.345337 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:45.345454 systemd[1]: kubelet.service: Consumed 121ms CPU time, 95M memory peak. Oct 13 00:02:45.347624 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:45.516372 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:45.528014 (kubelet)[2383]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 13 00:02:45.583326 kubelet[2383]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 00:02:45.583326 kubelet[2383]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 13 00:02:45.583326 kubelet[2383]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 00:02:45.583773 kubelet[2383]: I1013 00:02:45.583327 2383 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 13 00:02:46.660314 kubelet[2383]: I1013 00:02:46.660204 2383 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 13 00:02:46.660314 kubelet[2383]: I1013 00:02:46.660255 2383 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 13 00:02:46.661006 kubelet[2383]: I1013 00:02:46.660651 2383 server.go:954] "Client rotation is on, will bootstrap in background" Oct 13 00:02:46.696284 kubelet[2383]: E1013 00:02:46.696223 2383 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://49.13.69.12:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 49.13.69.12:6443: connect: connection refused" logger="UnhandledError" Oct 13 00:02:46.701440 kubelet[2383]: I1013 00:02:46.700535 2383 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 13 00:02:46.710068 kubelet[2383]: I1013 00:02:46.710024 2383 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 13 00:02:46.713498 kubelet[2383]: I1013 00:02:46.713458 2383 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 13 00:02:46.714956 kubelet[2383]: I1013 00:02:46.714847 2383 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 13 00:02:46.715265 kubelet[2383]: I1013 00:02:46.714937 2383 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459-1-0-3-bea61e6062","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 13 00:02:46.715421 kubelet[2383]: I1013 00:02:46.715332 2383 topology_manager.go:138] "Creating topology manager with none policy" Oct 13 00:02:46.715421 kubelet[2383]: I1013 00:02:46.715352 2383 container_manager_linux.go:304] "Creating device plugin manager" Oct 13 00:02:46.715697 kubelet[2383]: I1013 00:02:46.715644 2383 state_mem.go:36] "Initialized new in-memory state store" Oct 13 00:02:46.719282 kubelet[2383]: I1013 00:02:46.719206 2383 kubelet.go:446] "Attempting to sync node with API server" Oct 13 00:02:46.719282 kubelet[2383]: I1013 00:02:46.719251 2383 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 13 00:02:46.719282 kubelet[2383]: I1013 00:02:46.719285 2383 kubelet.go:352] "Adding apiserver pod source" Oct 13 00:02:46.719282 kubelet[2383]: I1013 00:02:46.719300 2383 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 13 00:02:46.723768 kubelet[2383]: W1013 00:02:46.723637 2383 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://49.13.69.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 49.13.69.12:6443: connect: connection refused Oct 13 00:02:46.723768 kubelet[2383]: E1013 00:02:46.723715 2383 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://49.13.69.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 49.13.69.12:6443: connect: connection refused" logger="UnhandledError" Oct 13 00:02:46.723922 kubelet[2383]: I1013 00:02:46.723810 2383 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 13 00:02:46.724613 kubelet[2383]: I1013 00:02:46.724516 2383 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 13 00:02:46.724720 kubelet[2383]: W1013 00:02:46.724704 2383 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Oct 13 00:02:46.726423 kubelet[2383]: I1013 00:02:46.726384 2383 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 13 00:02:46.726553 kubelet[2383]: I1013 00:02:46.726450 2383 server.go:1287] "Started kubelet" Oct 13 00:02:46.732442 kubelet[2383]: E1013 00:02:46.731688 2383 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://49.13.69.12:6443/api/v1/namespaces/default/events\": dial tcp 49.13.69.12:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4459-1-0-3-bea61e6062.186de40b56f551aa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4459-1-0-3-bea61e6062,UID:ci-4459-1-0-3-bea61e6062,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4459-1-0-3-bea61e6062,},FirstTimestamp:2025-10-13 00:02:46.726422954 +0000 UTC m=+1.191266673,LastTimestamp:2025-10-13 00:02:46.726422954 +0000 UTC m=+1.191266673,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459-1-0-3-bea61e6062,}" Oct 13 00:02:46.732442 kubelet[2383]: W1013 00:02:46.732049 2383 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://49.13.69.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4459-1-0-3-bea61e6062&limit=500&resourceVersion=0": dial tcp 49.13.69.12:6443: connect: connection refused Oct 13 00:02:46.732442 kubelet[2383]: E1013 00:02:46.732144 2383 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://49.13.69.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4459-1-0-3-bea61e6062&limit=500&resourceVersion=0\": dial tcp 49.13.69.12:6443: connect: connection refused" logger="UnhandledError" Oct 13 00:02:46.732442 kubelet[2383]: I1013 00:02:46.732249 2383 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 13 00:02:46.733827 kubelet[2383]: I1013 00:02:46.733744 2383 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 13 00:02:46.734277 kubelet[2383]: I1013 00:02:46.734243 2383 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 13 00:02:46.735939 kubelet[2383]: I1013 00:02:46.735915 2383 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 13 00:02:46.738843 kubelet[2383]: I1013 00:02:46.738741 2383 server.go:479] "Adding debug handlers to kubelet server" Oct 13 00:02:46.739973 kubelet[2383]: I1013 00:02:46.739926 2383 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 13 00:02:46.743881 kubelet[2383]: E1013 00:02:46.743342 2383 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4459-1-0-3-bea61e6062\" not found" Oct 13 00:02:46.743881 kubelet[2383]: I1013 00:02:46.743808 2383 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 13 00:02:46.744147 kubelet[2383]: I1013 00:02:46.744113 2383 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 13 00:02:46.744200 kubelet[2383]: I1013 00:02:46.744193 2383 reconciler.go:26] "Reconciler: start to sync state" Oct 13 00:02:46.744904 kubelet[2383]: W1013 00:02:46.744702 2383 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://49.13.69.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 49.13.69.12:6443: connect: connection refused Oct 13 00:02:46.744904 kubelet[2383]: E1013 00:02:46.744813 2383 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://49.13.69.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 49.13.69.12:6443: connect: connection refused" logger="UnhandledError" Oct 13 00:02:46.746674 kubelet[2383]: I1013 00:02:46.745319 2383 factory.go:221] Registration of the systemd container factory successfully Oct 13 00:02:46.746674 kubelet[2383]: I1013 00:02:46.745518 2383 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 13 00:02:46.746847 kubelet[2383]: E1013 00:02:46.746790 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://49.13.69.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-1-0-3-bea61e6062?timeout=10s\": dial tcp 49.13.69.12:6443: connect: connection refused" interval="200ms" Oct 13 00:02:46.747038 kubelet[2383]: E1013 00:02:46.747015 2383 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 13 00:02:46.747688 kubelet[2383]: I1013 00:02:46.747654 2383 factory.go:221] Registration of the containerd container factory successfully Oct 13 00:02:46.768792 kubelet[2383]: I1013 00:02:46.767876 2383 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 13 00:02:46.771567 kubelet[2383]: I1013 00:02:46.771505 2383 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 13 00:02:46.771567 kubelet[2383]: I1013 00:02:46.771559 2383 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 13 00:02:46.771739 kubelet[2383]: I1013 00:02:46.771591 2383 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 13 00:02:46.771739 kubelet[2383]: I1013 00:02:46.771609 2383 kubelet.go:2382] "Starting kubelet main sync loop" Oct 13 00:02:46.771739 kubelet[2383]: E1013 00:02:46.771687 2383 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 13 00:02:46.775517 kubelet[2383]: I1013 00:02:46.775490 2383 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 13 00:02:46.775650 kubelet[2383]: I1013 00:02:46.775638 2383 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 13 00:02:46.775738 kubelet[2383]: I1013 00:02:46.775729 2383 state_mem.go:36] "Initialized new in-memory state store" Oct 13 00:02:46.778599 kubelet[2383]: I1013 00:02:46.778555 2383 policy_none.go:49] "None policy: Start" Oct 13 00:02:46.779029 kubelet[2383]: I1013 00:02:46.778751 2383 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 13 00:02:46.779029 kubelet[2383]: I1013 00:02:46.778771 2383 state_mem.go:35] "Initializing new in-memory state store" Oct 13 00:02:46.781678 kubelet[2383]: W1013 00:02:46.781616 2383 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://49.13.69.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 49.13.69.12:6443: connect: connection refused Oct 13 00:02:46.781823 kubelet[2383]: E1013 00:02:46.781805 2383 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://49.13.69.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 49.13.69.12:6443: connect: connection refused" logger="UnhandledError" Oct 13 00:02:46.787838 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Oct 13 00:02:46.807043 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Oct 13 00:02:46.812793 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Oct 13 00:02:46.826436 kubelet[2383]: I1013 00:02:46.826342 2383 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 13 00:02:46.826755 kubelet[2383]: I1013 00:02:46.826667 2383 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 13 00:02:46.826755 kubelet[2383]: I1013 00:02:46.826698 2383 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 13 00:02:46.827715 kubelet[2383]: I1013 00:02:46.827685 2383 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 13 00:02:46.829398 kubelet[2383]: E1013 00:02:46.829041 2383 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 13 00:02:46.829398 kubelet[2383]: E1013 00:02:46.829184 2383 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4459-1-0-3-bea61e6062\" not found" Oct 13 00:02:46.903430 systemd[1]: Created slice kubepods-burstable-pod99dd9855356f30489a71fc4a98aba0fb.slice - libcontainer container kubepods-burstable-pod99dd9855356f30489a71fc4a98aba0fb.slice. Oct 13 00:02:46.904454 kernel: hrtimer: interrupt took 2379497 ns Oct 13 00:02:46.914814 kubelet[2383]: E1013 00:02:46.913621 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:46.922045 systemd[1]: Created slice kubepods-burstable-pod898173b5364a10961cbbcdf83f1fe100.slice - libcontainer container kubepods-burstable-pod898173b5364a10961cbbcdf83f1fe100.slice. Oct 13 00:02:46.927644 kubelet[2383]: E1013 00:02:46.927465 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:46.929855 systemd[1]: Created slice kubepods-burstable-podcb2952b2b29a7f47b55f00542af22ebf.slice - libcontainer container kubepods-burstable-podcb2952b2b29a7f47b55f00542af22ebf.slice. Oct 13 00:02:46.931265 kubelet[2383]: I1013 00:02:46.931214 2383 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:46.932133 kubelet[2383]: E1013 00:02:46.932102 2383 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://49.13.69.12:6443/api/v1/nodes\": dial tcp 49.13.69.12:6443: connect: connection refused" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:46.933505 kubelet[2383]: E1013 00:02:46.933476 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:46.947633 kubelet[2383]: E1013 00:02:46.947577 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://49.13.69.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-1-0-3-bea61e6062?timeout=10s\": dial tcp 49.13.69.12:6443: connect: connection refused" interval="400ms" Oct 13 00:02:47.045589 kubelet[2383]: I1013 00:02:47.045202 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-ca-certs\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045589 kubelet[2383]: I1013 00:02:47.045256 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045589 kubelet[2383]: I1013 00:02:47.045280 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/898173b5364a10961cbbcdf83f1fe100-kubeconfig\") pod \"kube-scheduler-ci-4459-1-0-3-bea61e6062\" (UID: \"898173b5364a10961cbbcdf83f1fe100\") " pod="kube-system/kube-scheduler-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045589 kubelet[2383]: I1013 00:02:47.045300 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/cb2952b2b29a7f47b55f00542af22ebf-k8s-certs\") pod \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" (UID: \"cb2952b2b29a7f47b55f00542af22ebf\") " pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045589 kubelet[2383]: I1013 00:02:47.045323 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-flexvolume-dir\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045887 kubelet[2383]: I1013 00:02:47.045341 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-k8s-certs\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045887 kubelet[2383]: I1013 00:02:47.045362 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-kubeconfig\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045887 kubelet[2383]: I1013 00:02:47.045381 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/cb2952b2b29a7f47b55f00542af22ebf-ca-certs\") pod \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" (UID: \"cb2952b2b29a7f47b55f00542af22ebf\") " pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.045887 kubelet[2383]: I1013 00:02:47.045432 2383 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/cb2952b2b29a7f47b55f00542af22ebf-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" (UID: \"cb2952b2b29a7f47b55f00542af22ebf\") " pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.135100 kubelet[2383]: I1013 00:02:47.134992 2383 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.135609 kubelet[2383]: E1013 00:02:47.135529 2383 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://49.13.69.12:6443/api/v1/nodes\": dial tcp 49.13.69.12:6443: connect: connection refused" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.218123 containerd[1544]: time="2025-10-13T00:02:47.217038357Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459-1-0-3-bea61e6062,Uid:99dd9855356f30489a71fc4a98aba0fb,Namespace:kube-system,Attempt:0,}" Oct 13 00:02:47.230488 containerd[1544]: time="2025-10-13T00:02:47.230289903Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459-1-0-3-bea61e6062,Uid:898173b5364a10961cbbcdf83f1fe100,Namespace:kube-system,Attempt:0,}" Oct 13 00:02:47.235444 containerd[1544]: time="2025-10-13T00:02:47.235237273Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459-1-0-3-bea61e6062,Uid:cb2952b2b29a7f47b55f00542af22ebf,Namespace:kube-system,Attempt:0,}" Oct 13 00:02:47.259392 containerd[1544]: time="2025-10-13T00:02:47.258530907Z" level=info msg="connecting to shim 67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226" address="unix:///run/containerd/s/89567c988c35bb5f53ce13c330eeefbd9f42c9d384d07e4e10c9be79a3a0f0f7" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:02:47.308649 containerd[1544]: time="2025-10-13T00:02:47.308061834Z" level=info msg="connecting to shim 2b92726a6062b9ee14e6e113c55763bb3ee3410a49bbeebf1823dcf41376fc38" address="unix:///run/containerd/s/65e4fe520ee6415cdc54f3465fa486765bf0ec2d87487b19f64d1fb6eebdfc79" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:02:47.312231 containerd[1544]: time="2025-10-13T00:02:47.312181667Z" level=info msg="connecting to shim 4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1" address="unix:///run/containerd/s/8acae749befb3ed7270aa02c8eb5a978b36c1cc07ff988806a9b33f48a499b3d" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:02:47.316733 systemd[1]: Started cri-containerd-67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226.scope - libcontainer container 67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226. Oct 13 00:02:47.348537 kubelet[2383]: E1013 00:02:47.348490 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://49.13.69.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-1-0-3-bea61e6062?timeout=10s\": dial tcp 49.13.69.12:6443: connect: connection refused" interval="800ms" Oct 13 00:02:47.368833 systemd[1]: Started cri-containerd-4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1.scope - libcontainer container 4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1. Oct 13 00:02:47.376136 systemd[1]: Started cri-containerd-2b92726a6062b9ee14e6e113c55763bb3ee3410a49bbeebf1823dcf41376fc38.scope - libcontainer container 2b92726a6062b9ee14e6e113c55763bb3ee3410a49bbeebf1823dcf41376fc38. Oct 13 00:02:47.394985 containerd[1544]: time="2025-10-13T00:02:47.394920257Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459-1-0-3-bea61e6062,Uid:99dd9855356f30489a71fc4a98aba0fb,Namespace:kube-system,Attempt:0,} returns sandbox id \"67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226\"" Oct 13 00:02:47.403303 containerd[1544]: time="2025-10-13T00:02:47.403239320Z" level=info msg="CreateContainer within sandbox \"67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Oct 13 00:02:47.419478 containerd[1544]: time="2025-10-13T00:02:47.417818711Z" level=info msg="Container e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:02:47.431113 containerd[1544]: time="2025-10-13T00:02:47.431046647Z" level=info msg="CreateContainer within sandbox \"67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5\"" Oct 13 00:02:47.433699 containerd[1544]: time="2025-10-13T00:02:47.433654233Z" level=info msg="StartContainer for \"e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5\"" Oct 13 00:02:47.447648 containerd[1544]: time="2025-10-13T00:02:47.447590251Z" level=info msg="connecting to shim e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5" address="unix:///run/containerd/s/89567c988c35bb5f53ce13c330eeefbd9f42c9d384d07e4e10c9be79a3a0f0f7" protocol=ttrpc version=3 Oct 13 00:02:47.454975 containerd[1544]: time="2025-10-13T00:02:47.454923866Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459-1-0-3-bea61e6062,Uid:898173b5364a10961cbbcdf83f1fe100,Namespace:kube-system,Attempt:0,} returns sandbox id \"4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1\"" Oct 13 00:02:47.461964 containerd[1544]: time="2025-10-13T00:02:47.461915166Z" level=info msg="CreateContainer within sandbox \"4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Oct 13 00:02:47.473713 containerd[1544]: time="2025-10-13T00:02:47.473591396Z" level=info msg="Container cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:02:47.476376 containerd[1544]: time="2025-10-13T00:02:47.476317756Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459-1-0-3-bea61e6062,Uid:cb2952b2b29a7f47b55f00542af22ebf,Namespace:kube-system,Attempt:0,} returns sandbox id \"2b92726a6062b9ee14e6e113c55763bb3ee3410a49bbeebf1823dcf41376fc38\"" Oct 13 00:02:47.478710 systemd[1]: Started cri-containerd-e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5.scope - libcontainer container e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5. Oct 13 00:02:47.484463 containerd[1544]: time="2025-10-13T00:02:47.483482614Z" level=info msg="CreateContainer within sandbox \"2b92726a6062b9ee14e6e113c55763bb3ee3410a49bbeebf1823dcf41376fc38\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Oct 13 00:02:47.499293 containerd[1544]: time="2025-10-13T00:02:47.499233378Z" level=info msg="CreateContainer within sandbox \"4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798\"" Oct 13 00:02:47.501635 containerd[1544]: time="2025-10-13T00:02:47.501559316Z" level=info msg="StartContainer for \"cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798\"" Oct 13 00:02:47.503075 containerd[1544]: time="2025-10-13T00:02:47.503002052Z" level=info msg="Container 7ae5a2969310a1a969b8f78811e04f0d829ca35d56ffdb1dfc33f912684be0c9: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:02:47.503183 containerd[1544]: time="2025-10-13T00:02:47.503026423Z" level=info msg="connecting to shim cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798" address="unix:///run/containerd/s/8acae749befb3ed7270aa02c8eb5a978b36c1cc07ff988806a9b33f48a499b3d" protocol=ttrpc version=3 Oct 13 00:02:47.515196 containerd[1544]: time="2025-10-13T00:02:47.515129527Z" level=info msg="CreateContainer within sandbox \"2b92726a6062b9ee14e6e113c55763bb3ee3410a49bbeebf1823dcf41376fc38\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"7ae5a2969310a1a969b8f78811e04f0d829ca35d56ffdb1dfc33f912684be0c9\"" Oct 13 00:02:47.518600 containerd[1544]: time="2025-10-13T00:02:47.518553925Z" level=info msg="StartContainer for \"7ae5a2969310a1a969b8f78811e04f0d829ca35d56ffdb1dfc33f912684be0c9\"" Oct 13 00:02:47.528956 containerd[1544]: time="2025-10-13T00:02:47.528852088Z" level=info msg="connecting to shim 7ae5a2969310a1a969b8f78811e04f0d829ca35d56ffdb1dfc33f912684be0c9" address="unix:///run/containerd/s/65e4fe520ee6415cdc54f3465fa486765bf0ec2d87487b19f64d1fb6eebdfc79" protocol=ttrpc version=3 Oct 13 00:02:47.536665 systemd[1]: Started cri-containerd-cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798.scope - libcontainer container cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798. Oct 13 00:02:47.540551 kubelet[2383]: I1013 00:02:47.540514 2383 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.546824 kubelet[2383]: E1013 00:02:47.546713 2383 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://49.13.69.12:6443/api/v1/nodes\": dial tcp 49.13.69.12:6443: connect: connection refused" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.557339 containerd[1544]: time="2025-10-13T00:02:47.557188976Z" level=info msg="StartContainer for \"e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5\" returns successfully" Oct 13 00:02:47.590684 systemd[1]: Started cri-containerd-7ae5a2969310a1a969b8f78811e04f0d829ca35d56ffdb1dfc33f912684be0c9.scope - libcontainer container 7ae5a2969310a1a969b8f78811e04f0d829ca35d56ffdb1dfc33f912684be0c9. Oct 13 00:02:47.638423 kubelet[2383]: W1013 00:02:47.638348 2383 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://49.13.69.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 49.13.69.12:6443: connect: connection refused Oct 13 00:02:47.638718 kubelet[2383]: E1013 00:02:47.638603 2383 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://49.13.69.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 49.13.69.12:6443: connect: connection refused" logger="UnhandledError" Oct 13 00:02:47.655277 containerd[1544]: time="2025-10-13T00:02:47.654565582Z" level=info msg="StartContainer for \"cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798\" returns successfully" Oct 13 00:02:47.668013 containerd[1544]: time="2025-10-13T00:02:47.667363283Z" level=info msg="StartContainer for \"7ae5a2969310a1a969b8f78811e04f0d829ca35d56ffdb1dfc33f912684be0c9\" returns successfully" Oct 13 00:02:47.782505 kubelet[2383]: E1013 00:02:47.782118 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.789844 kubelet[2383]: E1013 00:02:47.789442 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:47.790885 kubelet[2383]: E1013 00:02:47.790850 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:48.349504 kubelet[2383]: I1013 00:02:48.349433 2383 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:48.794259 kubelet[2383]: E1013 00:02:48.794141 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:48.794774 kubelet[2383]: E1013 00:02:48.794754 2383 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.020032 kubelet[2383]: E1013 00:02:50.019979 2383 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4459-1-0-3-bea61e6062\" not found" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.082212 kubelet[2383]: I1013 00:02:50.081933 2383 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.146847 kubelet[2383]: I1013 00:02:50.146797 2383 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.212511 kubelet[2383]: E1013 00:02:50.212458 2383 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459-1-0-3-bea61e6062\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.212511 kubelet[2383]: I1013 00:02:50.212504 2383 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.221008 kubelet[2383]: E1013 00:02:50.220693 2383 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.221008 kubelet[2383]: I1013 00:02:50.220749 2383 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.229891 kubelet[2383]: E1013 00:02:50.229813 2383 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.526589 kubelet[2383]: I1013 00:02:50.526550 2383 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.532427 kubelet[2383]: E1013 00:02:50.530794 2383 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:50.725101 kubelet[2383]: I1013 00:02:50.724833 2383 apiserver.go:52] "Watching apiserver" Oct 13 00:02:50.744852 kubelet[2383]: I1013 00:02:50.744802 2383 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 13 00:02:52.485997 systemd[1]: Reload requested from client PID 2651 ('systemctl') (unit session-7.scope)... Oct 13 00:02:52.486385 systemd[1]: Reloading... Oct 13 00:02:52.618461 zram_generator::config[2701]: No configuration found. Oct 13 00:02:52.830289 systemd[1]: Reloading finished in 343 ms. Oct 13 00:02:52.864279 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:52.884094 systemd[1]: kubelet.service: Deactivated successfully. Oct 13 00:02:52.884402 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:52.884504 systemd[1]: kubelet.service: Consumed 1.699s CPU time, 128M memory peak. Oct 13 00:02:52.887379 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 00:02:53.054707 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 00:02:53.066027 (kubelet)[2740]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 13 00:02:53.135241 kubelet[2740]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 00:02:53.135241 kubelet[2740]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 13 00:02:53.135241 kubelet[2740]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 00:02:53.136389 kubelet[2740]: I1013 00:02:53.135935 2740 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 13 00:02:53.148148 kubelet[2740]: I1013 00:02:53.148101 2740 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 13 00:02:53.149462 kubelet[2740]: I1013 00:02:53.148352 2740 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 13 00:02:53.149462 kubelet[2740]: I1013 00:02:53.148753 2740 server.go:954] "Client rotation is on, will bootstrap in background" Oct 13 00:02:53.152834 kubelet[2740]: I1013 00:02:53.152795 2740 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 13 00:02:53.156935 kubelet[2740]: I1013 00:02:53.156864 2740 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 13 00:02:53.162250 kubelet[2740]: I1013 00:02:53.162213 2740 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 13 00:02:53.165425 kubelet[2740]: I1013 00:02:53.165363 2740 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 13 00:02:53.165764 kubelet[2740]: I1013 00:02:53.165720 2740 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 13 00:02:53.165999 kubelet[2740]: I1013 00:02:53.165759 2740 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459-1-0-3-bea61e6062","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 13 00:02:53.165999 kubelet[2740]: I1013 00:02:53.165985 2740 topology_manager.go:138] "Creating topology manager with none policy" Oct 13 00:02:53.165999 kubelet[2740]: I1013 00:02:53.165995 2740 container_manager_linux.go:304] "Creating device plugin manager" Oct 13 00:02:53.166227 kubelet[2740]: I1013 00:02:53.166038 2740 state_mem.go:36] "Initialized new in-memory state store" Oct 13 00:02:53.166227 kubelet[2740]: I1013 00:02:53.166192 2740 kubelet.go:446] "Attempting to sync node with API server" Oct 13 00:02:53.166227 kubelet[2740]: I1013 00:02:53.166205 2740 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 13 00:02:53.166227 kubelet[2740]: I1013 00:02:53.166227 2740 kubelet.go:352] "Adding apiserver pod source" Oct 13 00:02:53.166866 kubelet[2740]: I1013 00:02:53.166837 2740 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 13 00:02:53.168732 kubelet[2740]: I1013 00:02:53.168697 2740 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 13 00:02:53.170012 kubelet[2740]: I1013 00:02:53.169980 2740 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 13 00:02:53.171356 kubelet[2740]: I1013 00:02:53.171323 2740 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 13 00:02:53.171470 kubelet[2740]: I1013 00:02:53.171371 2740 server.go:1287] "Started kubelet" Oct 13 00:02:53.176757 kubelet[2740]: I1013 00:02:53.176689 2740 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 13 00:02:53.178517 kubelet[2740]: I1013 00:02:53.178455 2740 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 13 00:02:53.182445 kubelet[2740]: I1013 00:02:53.181868 2740 server.go:479] "Adding debug handlers to kubelet server" Oct 13 00:02:53.185747 kubelet[2740]: I1013 00:02:53.185632 2740 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 13 00:02:53.186145 kubelet[2740]: I1013 00:02:53.186124 2740 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 13 00:02:53.187677 kubelet[2740]: I1013 00:02:53.187646 2740 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 13 00:02:53.188955 kubelet[2740]: E1013 00:02:53.188077 2740 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4459-1-0-3-bea61e6062\" not found" Oct 13 00:02:53.190896 kubelet[2740]: I1013 00:02:53.189997 2740 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 13 00:02:53.192050 kubelet[2740]: I1013 00:02:53.192018 2740 reconciler.go:26] "Reconciler: start to sync state" Oct 13 00:02:53.193051 kubelet[2740]: I1013 00:02:53.193016 2740 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 13 00:02:53.211990 kubelet[2740]: I1013 00:02:53.211949 2740 factory.go:221] Registration of the containerd container factory successfully Oct 13 00:02:53.211990 kubelet[2740]: I1013 00:02:53.211975 2740 factory.go:221] Registration of the systemd container factory successfully Oct 13 00:02:53.212129 kubelet[2740]: I1013 00:02:53.212094 2740 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 13 00:02:53.235385 kubelet[2740]: I1013 00:02:53.235040 2740 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 13 00:02:53.237490 kubelet[2740]: I1013 00:02:53.237258 2740 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 13 00:02:53.237490 kubelet[2740]: I1013 00:02:53.237472 2740 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 13 00:02:53.237780 kubelet[2740]: I1013 00:02:53.237499 2740 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 13 00:02:53.237780 kubelet[2740]: I1013 00:02:53.237506 2740 kubelet.go:2382] "Starting kubelet main sync loop" Oct 13 00:02:53.237780 kubelet[2740]: E1013 00:02:53.237545 2740 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 13 00:02:53.289245 kubelet[2740]: I1013 00:02:53.289216 2740 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 13 00:02:53.289463 kubelet[2740]: I1013 00:02:53.289448 2740 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 13 00:02:53.289532 kubelet[2740]: I1013 00:02:53.289523 2740 state_mem.go:36] "Initialized new in-memory state store" Oct 13 00:02:53.289805 kubelet[2740]: I1013 00:02:53.289785 2740 state_mem.go:88] "Updated default CPUSet" cpuSet="" Oct 13 00:02:53.289906 kubelet[2740]: I1013 00:02:53.289879 2740 state_mem.go:96] "Updated CPUSet assignments" assignments={} Oct 13 00:02:53.289959 kubelet[2740]: I1013 00:02:53.289951 2740 policy_none.go:49] "None policy: Start" Oct 13 00:02:53.290016 kubelet[2740]: I1013 00:02:53.290006 2740 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 13 00:02:53.290070 kubelet[2740]: I1013 00:02:53.290063 2740 state_mem.go:35] "Initializing new in-memory state store" Oct 13 00:02:53.290347 kubelet[2740]: I1013 00:02:53.290275 2740 state_mem.go:75] "Updated machine memory state" Oct 13 00:02:53.298294 kubelet[2740]: I1013 00:02:53.297972 2740 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 13 00:02:53.299538 kubelet[2740]: I1013 00:02:53.299277 2740 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 13 00:02:53.300328 kubelet[2740]: I1013 00:02:53.299706 2740 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 13 00:02:53.300865 kubelet[2740]: I1013 00:02:53.300753 2740 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 13 00:02:53.305304 kubelet[2740]: E1013 00:02:53.305273 2740 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 13 00:02:53.338125 kubelet[2740]: I1013 00:02:53.338088 2740 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.338655 kubelet[2740]: I1013 00:02:53.338097 2740 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.338655 kubelet[2740]: I1013 00:02:53.338155 2740 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.419956 kubelet[2740]: I1013 00:02:53.418486 2740 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.431355 kubelet[2740]: I1013 00:02:53.431298 2740 kubelet_node_status.go:124] "Node was previously registered" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.431499 kubelet[2740]: I1013 00:02:53.431437 2740 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493154 kubelet[2740]: I1013 00:02:53.493087 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-kubeconfig\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493154 kubelet[2740]: I1013 00:02:53.493141 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/898173b5364a10961cbbcdf83f1fe100-kubeconfig\") pod \"kube-scheduler-ci-4459-1-0-3-bea61e6062\" (UID: \"898173b5364a10961cbbcdf83f1fe100\") " pod="kube-system/kube-scheduler-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493331 kubelet[2740]: I1013 00:02:53.493167 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/cb2952b2b29a7f47b55f00542af22ebf-ca-certs\") pod \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" (UID: \"cb2952b2b29a7f47b55f00542af22ebf\") " pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493331 kubelet[2740]: I1013 00:02:53.493190 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/cb2952b2b29a7f47b55f00542af22ebf-k8s-certs\") pod \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" (UID: \"cb2952b2b29a7f47b55f00542af22ebf\") " pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493331 kubelet[2740]: I1013 00:02:53.493230 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/cb2952b2b29a7f47b55f00542af22ebf-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" (UID: \"cb2952b2b29a7f47b55f00542af22ebf\") " pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493331 kubelet[2740]: I1013 00:02:53.493252 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-ca-certs\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493331 kubelet[2740]: I1013 00:02:53.493268 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-k8s-certs\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493532 kubelet[2740]: I1013 00:02:53.493302 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-flexvolume-dir\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:53.493532 kubelet[2740]: I1013 00:02:53.493324 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/99dd9855356f30489a71fc4a98aba0fb-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459-1-0-3-bea61e6062\" (UID: \"99dd9855356f30489a71fc4a98aba0fb\") " pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:54.179687 kubelet[2740]: I1013 00:02:54.179642 2740 apiserver.go:52] "Watching apiserver" Oct 13 00:02:54.192261 kubelet[2740]: I1013 00:02:54.192214 2740 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 13 00:02:54.277171 kubelet[2740]: I1013 00:02:54.277131 2740 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:54.290422 kubelet[2740]: E1013 00:02:54.290363 2740 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459-1-0-3-bea61e6062\" already exists" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" Oct 13 00:02:54.321436 kubelet[2740]: I1013 00:02:54.319339 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4459-1-0-3-bea61e6062" podStartSLOduration=1.319316638 podStartE2EDuration="1.319316638s" podCreationTimestamp="2025-10-13 00:02:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 00:02:54.307969305 +0000 UTC m=+1.234832404" watchObservedRunningTime="2025-10-13 00:02:54.319316638 +0000 UTC m=+1.246179737" Oct 13 00:02:54.332364 kubelet[2740]: I1013 00:02:54.332279 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4459-1-0-3-bea61e6062" podStartSLOduration=1.332260784 podStartE2EDuration="1.332260784s" podCreationTimestamp="2025-10-13 00:02:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 00:02:54.320145334 +0000 UTC m=+1.247008433" watchObservedRunningTime="2025-10-13 00:02:54.332260784 +0000 UTC m=+1.259123843" Oct 13 00:02:54.348023 kubelet[2740]: I1013 00:02:54.347889 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4459-1-0-3-bea61e6062" podStartSLOduration=1.347866309 podStartE2EDuration="1.347866309s" podCreationTimestamp="2025-10-13 00:02:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 00:02:54.333212197 +0000 UTC m=+1.260075296" watchObservedRunningTime="2025-10-13 00:02:54.347866309 +0000 UTC m=+1.274729408" Oct 13 00:02:59.125377 kubelet[2740]: I1013 00:02:59.125140 2740 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Oct 13 00:02:59.126720 containerd[1544]: time="2025-10-13T00:02:59.126683738Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Oct 13 00:02:59.127607 kubelet[2740]: I1013 00:02:59.127583 2740 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Oct 13 00:03:00.119826 systemd[1]: Created slice kubepods-besteffort-pod1da493eb_dc32_46ca_b6e6_78bd3e9dc70b.slice - libcontainer container kubepods-besteffort-pod1da493eb_dc32_46ca_b6e6_78bd3e9dc70b.slice. Oct 13 00:03:00.242504 kubelet[2740]: I1013 00:03:00.242462 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1da493eb-dc32-46ca-b6e6-78bd3e9dc70b-xtables-lock\") pod \"kube-proxy-99xwn\" (UID: \"1da493eb-dc32-46ca-b6e6-78bd3e9dc70b\") " pod="kube-system/kube-proxy-99xwn" Oct 13 00:03:00.243334 kubelet[2740]: I1013 00:03:00.243251 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1da493eb-dc32-46ca-b6e6-78bd3e9dc70b-lib-modules\") pod \"kube-proxy-99xwn\" (UID: \"1da493eb-dc32-46ca-b6e6-78bd3e9dc70b\") " pod="kube-system/kube-proxy-99xwn" Oct 13 00:03:00.243334 kubelet[2740]: I1013 00:03:00.243302 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/1da493eb-dc32-46ca-b6e6-78bd3e9dc70b-kube-proxy\") pod \"kube-proxy-99xwn\" (UID: \"1da493eb-dc32-46ca-b6e6-78bd3e9dc70b\") " pod="kube-system/kube-proxy-99xwn" Oct 13 00:03:00.243673 kubelet[2740]: I1013 00:03:00.243650 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9bq2\" (UniqueName: \"kubernetes.io/projected/1da493eb-dc32-46ca-b6e6-78bd3e9dc70b-kube-api-access-j9bq2\") pod \"kube-proxy-99xwn\" (UID: \"1da493eb-dc32-46ca-b6e6-78bd3e9dc70b\") " pod="kube-system/kube-proxy-99xwn" Oct 13 00:03:00.249530 systemd[1]: Created slice kubepods-besteffort-pod244acb8d_edd1_4830_b6f7_d42ad708306f.slice - libcontainer container kubepods-besteffort-pod244acb8d_edd1_4830_b6f7_d42ad708306f.slice. Oct 13 00:03:00.345003 kubelet[2740]: I1013 00:03:00.344941 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/244acb8d-edd1-4830-b6f7-d42ad708306f-var-lib-calico\") pod \"tigera-operator-755d956888-zx6rz\" (UID: \"244acb8d-edd1-4830-b6f7-d42ad708306f\") " pod="tigera-operator/tigera-operator-755d956888-zx6rz" Oct 13 00:03:00.345302 kubelet[2740]: I1013 00:03:00.345276 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gswx6\" (UniqueName: \"kubernetes.io/projected/244acb8d-edd1-4830-b6f7-d42ad708306f-kube-api-access-gswx6\") pod \"tigera-operator-755d956888-zx6rz\" (UID: \"244acb8d-edd1-4830-b6f7-d42ad708306f\") " pod="tigera-operator/tigera-operator-755d956888-zx6rz" Oct 13 00:03:00.432257 containerd[1544]: time="2025-10-13T00:03:00.430847202Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-99xwn,Uid:1da493eb-dc32-46ca-b6e6-78bd3e9dc70b,Namespace:kube-system,Attempt:0,}" Oct 13 00:03:00.473739 containerd[1544]: time="2025-10-13T00:03:00.473689174Z" level=info msg="connecting to shim a15b802248bb694f16e875979f01b656a5332745650384528a6dcaca69b09e49" address="unix:///run/containerd/s/d8bd9b52c34221013752d6d9b625c917da58af9e919f3e01b21a6da4acecda1b" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:00.510672 systemd[1]: Started cri-containerd-a15b802248bb694f16e875979f01b656a5332745650384528a6dcaca69b09e49.scope - libcontainer container a15b802248bb694f16e875979f01b656a5332745650384528a6dcaca69b09e49. Oct 13 00:03:00.542130 containerd[1544]: time="2025-10-13T00:03:00.542079741Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-99xwn,Uid:1da493eb-dc32-46ca-b6e6-78bd3e9dc70b,Namespace:kube-system,Attempt:0,} returns sandbox id \"a15b802248bb694f16e875979f01b656a5332745650384528a6dcaca69b09e49\"" Oct 13 00:03:00.548154 containerd[1544]: time="2025-10-13T00:03:00.548113061Z" level=info msg="CreateContainer within sandbox \"a15b802248bb694f16e875979f01b656a5332745650384528a6dcaca69b09e49\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Oct 13 00:03:00.557873 containerd[1544]: time="2025-10-13T00:03:00.557821688Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-zx6rz,Uid:244acb8d-edd1-4830-b6f7-d42ad708306f,Namespace:tigera-operator,Attempt:0,}" Oct 13 00:03:00.567449 containerd[1544]: time="2025-10-13T00:03:00.565873382Z" level=info msg="Container 5d9cfac778667aa1f712b71e868dcc91e53bd14ac3c0c8cdf8cecb5895de31cd: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:00.580118 containerd[1544]: time="2025-10-13T00:03:00.580019769Z" level=info msg="CreateContainer within sandbox \"a15b802248bb694f16e875979f01b656a5332745650384528a6dcaca69b09e49\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"5d9cfac778667aa1f712b71e868dcc91e53bd14ac3c0c8cdf8cecb5895de31cd\"" Oct 13 00:03:00.583324 containerd[1544]: time="2025-10-13T00:03:00.581832777Z" level=info msg="StartContainer for \"5d9cfac778667aa1f712b71e868dcc91e53bd14ac3c0c8cdf8cecb5895de31cd\"" Oct 13 00:03:00.597149 containerd[1544]: time="2025-10-13T00:03:00.597008996Z" level=info msg="connecting to shim 5d9cfac778667aa1f712b71e868dcc91e53bd14ac3c0c8cdf8cecb5895de31cd" address="unix:///run/containerd/s/d8bd9b52c34221013752d6d9b625c917da58af9e919f3e01b21a6da4acecda1b" protocol=ttrpc version=3 Oct 13 00:03:00.600963 containerd[1544]: time="2025-10-13T00:03:00.600897752Z" level=info msg="connecting to shim 878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982" address="unix:///run/containerd/s/b1102b39bd6fa2999ccdc54d7af04c1e31b62d5f2bb1b27426ed52f81abd3d6f" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:00.626643 systemd[1]: Started cri-containerd-5d9cfac778667aa1f712b71e868dcc91e53bd14ac3c0c8cdf8cecb5895de31cd.scope - libcontainer container 5d9cfac778667aa1f712b71e868dcc91e53bd14ac3c0c8cdf8cecb5895de31cd. Oct 13 00:03:00.635625 systemd[1]: Started cri-containerd-878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982.scope - libcontainer container 878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982. Oct 13 00:03:00.704695 containerd[1544]: time="2025-10-13T00:03:00.704272721Z" level=info msg="StartContainer for \"5d9cfac778667aa1f712b71e868dcc91e53bd14ac3c0c8cdf8cecb5895de31cd\" returns successfully" Oct 13 00:03:00.707641 containerd[1544]: time="2025-10-13T00:03:00.707590948Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-zx6rz,Uid:244acb8d-edd1-4830-b6f7-d42ad708306f,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982\"" Oct 13 00:03:00.711582 containerd[1544]: time="2025-10-13T00:03:00.711354356Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Oct 13 00:03:02.378349 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2758136467.mount: Deactivated successfully. Oct 13 00:03:02.810719 containerd[1544]: time="2025-10-13T00:03:02.810439893Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:02.812714 containerd[1544]: time="2025-10-13T00:03:02.812654945Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=22152365" Oct 13 00:03:02.814785 containerd[1544]: time="2025-10-13T00:03:02.814092758Z" level=info msg="ImageCreate event name:\"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:02.818142 containerd[1544]: time="2025-10-13T00:03:02.818098176Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:02.818637 containerd[1544]: time="2025-10-13T00:03:02.818595518Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"22148360\" in 2.106823067s" Oct 13 00:03:02.818637 containerd[1544]: time="2025-10-13T00:03:02.818635486Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\"" Oct 13 00:03:02.825372 containerd[1544]: time="2025-10-13T00:03:02.825318850Z" level=info msg="CreateContainer within sandbox \"878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Oct 13 00:03:02.834264 kubelet[2740]: I1013 00:03:02.834165 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-99xwn" podStartSLOduration=2.8341407309999997 podStartE2EDuration="2.834140731s" podCreationTimestamp="2025-10-13 00:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 00:03:01.314831245 +0000 UTC m=+8.241694344" watchObservedRunningTime="2025-10-13 00:03:02.834140731 +0000 UTC m=+9.761003830" Oct 13 00:03:02.846499 containerd[1544]: time="2025-10-13T00:03:02.845216392Z" level=info msg="Container 5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:02.855224 containerd[1544]: time="2025-10-13T00:03:02.855177505Z" level=info msg="CreateContainer within sandbox \"878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7\"" Oct 13 00:03:02.857792 containerd[1544]: time="2025-10-13T00:03:02.857731986Z" level=info msg="StartContainer for \"5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7\"" Oct 13 00:03:02.859493 containerd[1544]: time="2025-10-13T00:03:02.859434174Z" level=info msg="connecting to shim 5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7" address="unix:///run/containerd/s/b1102b39bd6fa2999ccdc54d7af04c1e31b62d5f2bb1b27426ed52f81abd3d6f" protocol=ttrpc version=3 Oct 13 00:03:02.885857 systemd[1]: Started cri-containerd-5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7.scope - libcontainer container 5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7. Oct 13 00:03:02.928108 containerd[1544]: time="2025-10-13T00:03:02.928056542Z" level=info msg="StartContainer for \"5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7\" returns successfully" Oct 13 00:03:03.355829 kubelet[2740]: I1013 00:03:03.355324 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-zx6rz" podStartSLOduration=1.245279895 podStartE2EDuration="3.355287005s" podCreationTimestamp="2025-10-13 00:03:00 +0000 UTC" firstStartedPulling="2025-10-13 00:03:00.710225102 +0000 UTC m=+7.637088201" lastFinishedPulling="2025-10-13 00:03:02.820232212 +0000 UTC m=+9.747095311" observedRunningTime="2025-10-13 00:03:03.35525716 +0000 UTC m=+10.282120219" watchObservedRunningTime="2025-10-13 00:03:03.355287005 +0000 UTC m=+10.282150064" Oct 13 00:03:09.366174 sudo[1818]: pam_unix(sudo:session): session closed for user root Oct 13 00:03:09.523266 sshd[1803]: Connection closed by 139.178.89.65 port 47278 Oct 13 00:03:09.523669 sshd-session[1800]: pam_unix(sshd:session): session closed for user core Oct 13 00:03:09.531810 systemd-logind[1521]: Session 7 logged out. Waiting for processes to exit. Oct 13 00:03:09.532004 systemd[1]: sshd@6-49.13.69.12:22-139.178.89.65:47278.service: Deactivated successfully. Oct 13 00:03:09.541315 systemd[1]: session-7.scope: Deactivated successfully. Oct 13 00:03:09.542260 systemd[1]: session-7.scope: Consumed 7.826s CPU time, 220.3M memory peak. Oct 13 00:03:09.544989 systemd-logind[1521]: Removed session 7. Oct 13 00:03:16.144784 systemd[1]: Created slice kubepods-besteffort-podd282594b_b441_4a90_b390_95ec7801f93b.slice - libcontainer container kubepods-besteffort-podd282594b_b441_4a90_b390_95ec7801f93b.slice. Oct 13 00:03:16.246630 kubelet[2740]: I1013 00:03:16.246570 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d282594b-b441-4a90-b390-95ec7801f93b-tigera-ca-bundle\") pod \"calico-typha-5cbf8867fd-7ll5n\" (UID: \"d282594b-b441-4a90-b390-95ec7801f93b\") " pod="calico-system/calico-typha-5cbf8867fd-7ll5n" Oct 13 00:03:16.246630 kubelet[2740]: I1013 00:03:16.246629 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cdzb\" (UniqueName: \"kubernetes.io/projected/d282594b-b441-4a90-b390-95ec7801f93b-kube-api-access-7cdzb\") pod \"calico-typha-5cbf8867fd-7ll5n\" (UID: \"d282594b-b441-4a90-b390-95ec7801f93b\") " pod="calico-system/calico-typha-5cbf8867fd-7ll5n" Oct 13 00:03:16.247138 kubelet[2740]: I1013 00:03:16.246653 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/d282594b-b441-4a90-b390-95ec7801f93b-typha-certs\") pod \"calico-typha-5cbf8867fd-7ll5n\" (UID: \"d282594b-b441-4a90-b390-95ec7801f93b\") " pod="calico-system/calico-typha-5cbf8867fd-7ll5n" Oct 13 00:03:16.304716 systemd[1]: Created slice kubepods-besteffort-pode9b13ba8_2978_476f_bf8f_b3ec2804c3b3.slice - libcontainer container kubepods-besteffort-pode9b13ba8_2978_476f_bf8f_b3ec2804c3b3.slice. Oct 13 00:03:16.347846 kubelet[2740]: I1013 00:03:16.347492 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-cni-bin-dir\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.347846 kubelet[2740]: I1013 00:03:16.347535 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-lib-modules\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.347846 kubelet[2740]: I1013 00:03:16.347551 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-policysync\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.347846 kubelet[2740]: I1013 00:03:16.347567 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-xtables-lock\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.347846 kubelet[2740]: I1013 00:03:16.347583 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-cni-net-dir\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.348118 kubelet[2740]: I1013 00:03:16.347597 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-node-certs\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.348118 kubelet[2740]: I1013 00:03:16.347636 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-cni-log-dir\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.348118 kubelet[2740]: I1013 00:03:16.347652 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-flexvol-driver-host\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.348118 kubelet[2740]: I1013 00:03:16.347669 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-var-lib-calico\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.348118 kubelet[2740]: I1013 00:03:16.347686 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-var-run-calico\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.348225 kubelet[2740]: I1013 00:03:16.347703 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-tigera-ca-bundle\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.348225 kubelet[2740]: I1013 00:03:16.347720 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l652q\" (UniqueName: \"kubernetes.io/projected/e9b13ba8-2978-476f-bf8f-b3ec2804c3b3-kube-api-access-l652q\") pod \"calico-node-hkq4m\" (UID: \"e9b13ba8-2978-476f-bf8f-b3ec2804c3b3\") " pod="calico-system/calico-node-hkq4m" Oct 13 00:03:16.437653 kubelet[2740]: E1013 00:03:16.436644 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mtmkk" podUID="c82ba31d-2f24-4bf3-a00d-09a349524d1c" Oct 13 00:03:16.448057 kubelet[2740]: I1013 00:03:16.448008 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/c82ba31d-2f24-4bf3-a00d-09a349524d1c-varrun\") pod \"csi-node-driver-mtmkk\" (UID: \"c82ba31d-2f24-4bf3-a00d-09a349524d1c\") " pod="calico-system/csi-node-driver-mtmkk" Oct 13 00:03:16.448482 kubelet[2740]: I1013 00:03:16.448452 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c82ba31d-2f24-4bf3-a00d-09a349524d1c-socket-dir\") pod \"csi-node-driver-mtmkk\" (UID: \"c82ba31d-2f24-4bf3-a00d-09a349524d1c\") " pod="calico-system/csi-node-driver-mtmkk" Oct 13 00:03:16.448553 kubelet[2740]: I1013 00:03:16.448529 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c82ba31d-2f24-4bf3-a00d-09a349524d1c-kubelet-dir\") pod \"csi-node-driver-mtmkk\" (UID: \"c82ba31d-2f24-4bf3-a00d-09a349524d1c\") " pod="calico-system/csi-node-driver-mtmkk" Oct 13 00:03:16.448767 kubelet[2740]: I1013 00:03:16.448556 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7z8t\" (UniqueName: \"kubernetes.io/projected/c82ba31d-2f24-4bf3-a00d-09a349524d1c-kube-api-access-r7z8t\") pod \"csi-node-driver-mtmkk\" (UID: \"c82ba31d-2f24-4bf3-a00d-09a349524d1c\") " pod="calico-system/csi-node-driver-mtmkk" Oct 13 00:03:16.448767 kubelet[2740]: I1013 00:03:16.448656 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c82ba31d-2f24-4bf3-a00d-09a349524d1c-registration-dir\") pod \"csi-node-driver-mtmkk\" (UID: \"c82ba31d-2f24-4bf3-a00d-09a349524d1c\") " pod="calico-system/csi-node-driver-mtmkk" Oct 13 00:03:16.454429 containerd[1544]: time="2025-10-13T00:03:16.453135590Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5cbf8867fd-7ll5n,Uid:d282594b-b441-4a90-b390-95ec7801f93b,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:16.460558 kubelet[2740]: E1013 00:03:16.460504 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.460558 kubelet[2740]: W1013 00:03:16.460536 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.460558 kubelet[2740]: E1013 00:03:16.460568 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.497630 kubelet[2740]: E1013 00:03:16.497590 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.497630 kubelet[2740]: W1013 00:03:16.497618 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.497842 kubelet[2740]: E1013 00:03:16.497644 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.503161 containerd[1544]: time="2025-10-13T00:03:16.502804752Z" level=info msg="connecting to shim 8ff56aac8e2748ea3e676f4c85ecd46e09c62585e8cd5139da32dfcafdd6f833" address="unix:///run/containerd/s/a62148604eb653d8639b6720f5a453b3000334872ed830cd455793608d2a3520" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:16.547651 systemd[1]: Started cri-containerd-8ff56aac8e2748ea3e676f4c85ecd46e09c62585e8cd5139da32dfcafdd6f833.scope - libcontainer container 8ff56aac8e2748ea3e676f4c85ecd46e09c62585e8cd5139da32dfcafdd6f833. Oct 13 00:03:16.551499 kubelet[2740]: E1013 00:03:16.551454 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.552227 kubelet[2740]: W1013 00:03:16.551748 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.552227 kubelet[2740]: E1013 00:03:16.551782 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.553516 kubelet[2740]: E1013 00:03:16.553490 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.553806 kubelet[2740]: W1013 00:03:16.553709 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.553806 kubelet[2740]: E1013 00:03:16.553746 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.554450 kubelet[2740]: E1013 00:03:16.554426 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.554730 kubelet[2740]: W1013 00:03:16.554597 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.554730 kubelet[2740]: E1013 00:03:16.554647 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.555146 kubelet[2740]: E1013 00:03:16.555127 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.555146 kubelet[2740]: W1013 00:03:16.555181 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.555146 kubelet[2740]: E1013 00:03:16.555204 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.555703 kubelet[2740]: E1013 00:03:16.555366 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.555703 kubelet[2740]: W1013 00:03:16.555378 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.555703 kubelet[2740]: E1013 00:03:16.555392 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.555703 kubelet[2740]: E1013 00:03:16.555618 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.555703 kubelet[2740]: W1013 00:03:16.555628 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.555703 kubelet[2740]: E1013 00:03:16.555642 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.556006 kubelet[2740]: E1013 00:03:16.555876 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.556006 kubelet[2740]: W1013 00:03:16.555897 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.556006 kubelet[2740]: E1013 00:03:16.555910 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.556263 kubelet[2740]: E1013 00:03:16.556075 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.556263 kubelet[2740]: W1013 00:03:16.556085 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.556263 kubelet[2740]: E1013 00:03:16.556103 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.557134 kubelet[2740]: E1013 00:03:16.557109 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.557291 kubelet[2740]: W1013 00:03:16.557132 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.557291 kubelet[2740]: E1013 00:03:16.557179 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.557589 kubelet[2740]: E1013 00:03:16.557569 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.557589 kubelet[2740]: W1013 00:03:16.557585 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.557589 kubelet[2740]: E1013 00:03:16.557604 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.558598 kubelet[2740]: E1013 00:03:16.558575 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.558598 kubelet[2740]: W1013 00:03:16.558594 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.558710 kubelet[2740]: E1013 00:03:16.558615 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.558887 kubelet[2740]: E1013 00:03:16.558871 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.558887 kubelet[2740]: W1013 00:03:16.558885 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.559099 kubelet[2740]: E1013 00:03:16.558990 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.559099 kubelet[2740]: E1013 00:03:16.559077 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.559099 kubelet[2740]: W1013 00:03:16.559084 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.559217 kubelet[2740]: E1013 00:03:16.559202 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.559301 kubelet[2740]: E1013 00:03:16.559209 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.559372 kubelet[2740]: W1013 00:03:16.559360 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.559531 kubelet[2740]: E1013 00:03:16.559504 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.560153 kubelet[2740]: E1013 00:03:16.560130 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.560417 kubelet[2740]: W1013 00:03:16.560304 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.560417 kubelet[2740]: E1013 00:03:16.560344 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.560786 kubelet[2740]: E1013 00:03:16.560679 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.560953 kubelet[2740]: W1013 00:03:16.560873 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.560953 kubelet[2740]: E1013 00:03:16.560944 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.561451 kubelet[2740]: E1013 00:03:16.561352 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.561451 kubelet[2740]: W1013 00:03:16.561365 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.561451 kubelet[2740]: E1013 00:03:16.561430 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.562060 kubelet[2740]: E1013 00:03:16.561833 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.562060 kubelet[2740]: W1013 00:03:16.561855 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.562060 kubelet[2740]: E1013 00:03:16.561898 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.562438 kubelet[2740]: E1013 00:03:16.562401 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.562795 kubelet[2740]: W1013 00:03:16.562509 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.562795 kubelet[2740]: E1013 00:03:16.562576 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.563282 kubelet[2740]: E1013 00:03:16.563220 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.563603 kubelet[2740]: W1013 00:03:16.563359 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.563603 kubelet[2740]: E1013 00:03:16.563438 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.563988 kubelet[2740]: E1013 00:03:16.563890 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.564212 kubelet[2740]: W1013 00:03:16.564061 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.564212 kubelet[2740]: E1013 00:03:16.564124 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.564533 kubelet[2740]: E1013 00:03:16.564511 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.564700 kubelet[2740]: W1013 00:03:16.564679 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.565128 kubelet[2740]: E1013 00:03:16.565106 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.565521 kubelet[2740]: E1013 00:03:16.565500 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.565521 kubelet[2740]: W1013 00:03:16.565517 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.565600 kubelet[2740]: E1013 00:03:16.565534 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.566052 kubelet[2740]: E1013 00:03:16.566008 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.566052 kubelet[2740]: W1013 00:03:16.566047 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.566164 kubelet[2740]: E1013 00:03:16.566071 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.567868 kubelet[2740]: E1013 00:03:16.567577 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.567868 kubelet[2740]: W1013 00:03:16.567608 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.567868 kubelet[2740]: E1013 00:03:16.567629 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.592579 kubelet[2740]: E1013 00:03:16.592546 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:16.592579 kubelet[2740]: W1013 00:03:16.592571 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:16.592736 kubelet[2740]: E1013 00:03:16.592593 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:16.610788 containerd[1544]: time="2025-10-13T00:03:16.610742404Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-hkq4m,Uid:e9b13ba8-2978-476f-bf8f-b3ec2804c3b3,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:16.656694 containerd[1544]: time="2025-10-13T00:03:16.656609097Z" level=info msg="connecting to shim 80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9" address="unix:///run/containerd/s/e38ca4d4b5fc773c56eca9a14c886710d14e02ca1e22b19e832656fe9232de0b" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:16.686514 containerd[1544]: time="2025-10-13T00:03:16.685633250Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5cbf8867fd-7ll5n,Uid:d282594b-b441-4a90-b390-95ec7801f93b,Namespace:calico-system,Attempt:0,} returns sandbox id \"8ff56aac8e2748ea3e676f4c85ecd46e09c62585e8cd5139da32dfcafdd6f833\"" Oct 13 00:03:16.691035 containerd[1544]: time="2025-10-13T00:03:16.690615812Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Oct 13 00:03:16.718714 systemd[1]: Started cri-containerd-80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9.scope - libcontainer container 80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9. Oct 13 00:03:16.788621 containerd[1544]: time="2025-10-13T00:03:16.788583861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-hkq4m,Uid:e9b13ba8-2978-476f-bf8f-b3ec2804c3b3,Namespace:calico-system,Attempt:0,} returns sandbox id \"80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9\"" Oct 13 00:03:18.238463 kubelet[2740]: E1013 00:03:18.238262 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mtmkk" podUID="c82ba31d-2f24-4bf3-a00d-09a349524d1c" Oct 13 00:03:18.439974 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3918281179.mount: Deactivated successfully. Oct 13 00:03:19.430997 containerd[1544]: time="2025-10-13T00:03:19.430834953Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:19.433097 containerd[1544]: time="2025-10-13T00:03:19.432984772Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=33105775" Oct 13 00:03:19.434292 containerd[1544]: time="2025-10-13T00:03:19.434193695Z" level=info msg="ImageCreate event name:\"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:19.437507 containerd[1544]: time="2025-10-13T00:03:19.437442466Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:19.438803 containerd[1544]: time="2025-10-13T00:03:19.438725877Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"33105629\" in 2.748063659s" Oct 13 00:03:19.438803 containerd[1544]: time="2025-10-13T00:03:19.438768881Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\"" Oct 13 00:03:19.440504 containerd[1544]: time="2025-10-13T00:03:19.440248952Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Oct 13 00:03:19.456782 containerd[1544]: time="2025-10-13T00:03:19.456744031Z" level=info msg="CreateContainer within sandbox \"8ff56aac8e2748ea3e676f4c85ecd46e09c62585e8cd5139da32dfcafdd6f833\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Oct 13 00:03:19.473525 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3871638433.mount: Deactivated successfully. Oct 13 00:03:19.474503 containerd[1544]: time="2025-10-13T00:03:19.474454675Z" level=info msg="Container 87b14ee7081f0aca1665e068b81046fb3ff3c529b76478804687851d6473b2de: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:19.486727 containerd[1544]: time="2025-10-13T00:03:19.486605672Z" level=info msg="CreateContainer within sandbox \"8ff56aac8e2748ea3e676f4c85ecd46e09c62585e8cd5139da32dfcafdd6f833\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"87b14ee7081f0aca1665e068b81046fb3ff3c529b76478804687851d6473b2de\"" Oct 13 00:03:19.487835 containerd[1544]: time="2025-10-13T00:03:19.487694583Z" level=info msg="StartContainer for \"87b14ee7081f0aca1665e068b81046fb3ff3c529b76478804687851d6473b2de\"" Oct 13 00:03:19.489986 containerd[1544]: time="2025-10-13T00:03:19.489909368Z" level=info msg="connecting to shim 87b14ee7081f0aca1665e068b81046fb3ff3c529b76478804687851d6473b2de" address="unix:///run/containerd/s/a62148604eb653d8639b6720f5a453b3000334872ed830cd455793608d2a3520" protocol=ttrpc version=3 Oct 13 00:03:19.518077 systemd[1]: Started cri-containerd-87b14ee7081f0aca1665e068b81046fb3ff3c529b76478804687851d6473b2de.scope - libcontainer container 87b14ee7081f0aca1665e068b81046fb3ff3c529b76478804687851d6473b2de. Oct 13 00:03:19.573391 containerd[1544]: time="2025-10-13T00:03:19.573349545Z" level=info msg="StartContainer for \"87b14ee7081f0aca1665e068b81046fb3ff3c529b76478804687851d6473b2de\" returns successfully" Oct 13 00:03:20.238542 kubelet[2740]: E1013 00:03:20.238432 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mtmkk" podUID="c82ba31d-2f24-4bf3-a00d-09a349524d1c" Oct 13 00:03:20.382031 kubelet[2740]: E1013 00:03:20.381984 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.382031 kubelet[2740]: W1013 00:03:20.382022 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.382201 kubelet[2740]: E1013 00:03:20.382048 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383025 kubelet[2740]: E1013 00:03:20.382228 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383025 kubelet[2740]: W1013 00:03:20.382265 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383025 kubelet[2740]: E1013 00:03:20.382345 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383025 kubelet[2740]: E1013 00:03:20.382542 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383025 kubelet[2740]: W1013 00:03:20.382552 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383025 kubelet[2740]: E1013 00:03:20.382561 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383025 kubelet[2740]: E1013 00:03:20.382714 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383025 kubelet[2740]: W1013 00:03:20.382722 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383025 kubelet[2740]: E1013 00:03:20.382730 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383025 kubelet[2740]: E1013 00:03:20.382930 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383453 kubelet[2740]: W1013 00:03:20.382940 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383453 kubelet[2740]: E1013 00:03:20.382949 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383453 kubelet[2740]: E1013 00:03:20.383113 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383453 kubelet[2740]: W1013 00:03:20.383122 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383453 kubelet[2740]: E1013 00:03:20.383130 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383453 kubelet[2740]: E1013 00:03:20.383281 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383453 kubelet[2740]: W1013 00:03:20.383288 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383453 kubelet[2740]: E1013 00:03:20.383324 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383643 kubelet[2740]: E1013 00:03:20.383490 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383643 kubelet[2740]: W1013 00:03:20.383498 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383643 kubelet[2740]: E1013 00:03:20.383506 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383774 kubelet[2740]: E1013 00:03:20.383660 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383774 kubelet[2740]: W1013 00:03:20.383667 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383774 kubelet[2740]: E1013 00:03:20.383675 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383867 kubelet[2740]: E1013 00:03:20.383811 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383867 kubelet[2740]: W1013 00:03:20.383819 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383867 kubelet[2740]: E1013 00:03:20.383826 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.383936 kubelet[2740]: E1013 00:03:20.383931 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.383963 kubelet[2740]: W1013 00:03:20.383937 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.383963 kubelet[2740]: E1013 00:03:20.383955 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.384072 kubelet[2740]: E1013 00:03:20.384063 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.384072 kubelet[2740]: W1013 00:03:20.384071 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.384125 kubelet[2740]: E1013 00:03:20.384078 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.384202 kubelet[2740]: E1013 00:03:20.384193 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.384202 kubelet[2740]: W1013 00:03:20.384201 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.384356 kubelet[2740]: E1013 00:03:20.384211 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.384594 kubelet[2740]: E1013 00:03:20.384519 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.384715 kubelet[2740]: W1013 00:03:20.384694 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.384747 kubelet[2740]: E1013 00:03:20.384717 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.385504 kubelet[2740]: E1013 00:03:20.385469 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.385504 kubelet[2740]: W1013 00:03:20.385490 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.385504 kubelet[2740]: E1013 00:03:20.385503 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.386940 kubelet[2740]: E1013 00:03:20.385834 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.386940 kubelet[2740]: W1013 00:03:20.385845 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.386940 kubelet[2740]: E1013 00:03:20.385870 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.386940 kubelet[2740]: E1013 00:03:20.386050 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.386940 kubelet[2740]: W1013 00:03:20.386059 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.386940 kubelet[2740]: E1013 00:03:20.386067 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.386940 kubelet[2740]: E1013 00:03:20.386230 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.386940 kubelet[2740]: W1013 00:03:20.386239 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.386940 kubelet[2740]: E1013 00:03:20.386249 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.386940 kubelet[2740]: E1013 00:03:20.386561 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.387156 kubelet[2740]: W1013 00:03:20.386574 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.387156 kubelet[2740]: E1013 00:03:20.386585 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.387156 kubelet[2740]: E1013 00:03:20.386736 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.387156 kubelet[2740]: W1013 00:03:20.386745 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.387156 kubelet[2740]: E1013 00:03:20.386769 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.387156 kubelet[2740]: E1013 00:03:20.386885 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.387156 kubelet[2740]: W1013 00:03:20.386898 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.387156 kubelet[2740]: E1013 00:03:20.386906 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.387156 kubelet[2740]: E1013 00:03:20.387100 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.387156 kubelet[2740]: W1013 00:03:20.387108 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.387429 kubelet[2740]: E1013 00:03:20.387117 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.387611 kubelet[2740]: E1013 00:03:20.387555 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.387611 kubelet[2740]: W1013 00:03:20.387594 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.387611 kubelet[2740]: E1013 00:03:20.387609 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.387808 kubelet[2740]: E1013 00:03:20.387790 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.387808 kubelet[2740]: W1013 00:03:20.387805 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.387869 kubelet[2740]: E1013 00:03:20.387829 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.388194 kubelet[2740]: E1013 00:03:20.387957 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.388194 kubelet[2740]: W1013 00:03:20.387970 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.388194 kubelet[2740]: E1013 00:03:20.387995 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.388194 kubelet[2740]: E1013 00:03:20.388154 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.388194 kubelet[2740]: W1013 00:03:20.388162 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.388194 kubelet[2740]: E1013 00:03:20.388178 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.390184 kubelet[2740]: E1013 00:03:20.388366 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.390184 kubelet[2740]: W1013 00:03:20.388375 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.390184 kubelet[2740]: E1013 00:03:20.388455 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.390481 kubelet[2740]: E1013 00:03:20.390461 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.390662 kubelet[2740]: W1013 00:03:20.390557 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.390662 kubelet[2740]: E1013 00:03:20.390616 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.391040 kubelet[2740]: E1013 00:03:20.391026 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.391197 kubelet[2740]: W1013 00:03:20.391183 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.391284 kubelet[2740]: E1013 00:03:20.391262 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.391551 kubelet[2740]: E1013 00:03:20.391534 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.393492 kubelet[2740]: W1013 00:03:20.393454 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.393631 kubelet[2740]: E1013 00:03:20.393617 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.393856 kubelet[2740]: E1013 00:03:20.393839 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.393900 kubelet[2740]: W1013 00:03:20.393854 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.393900 kubelet[2740]: E1013 00:03:20.393875 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.394056 kubelet[2740]: E1013 00:03:20.394045 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.394056 kubelet[2740]: W1013 00:03:20.394055 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.394114 kubelet[2740]: E1013 00:03:20.394064 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:20.394725 kubelet[2740]: E1013 00:03:20.394698 2740 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 00:03:20.394725 kubelet[2740]: W1013 00:03:20.394722 2740 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 00:03:20.394840 kubelet[2740]: E1013 00:03:20.394741 2740 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 00:03:21.146475 containerd[1544]: time="2025-10-13T00:03:21.146255396Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:21.148022 containerd[1544]: time="2025-10-13T00:03:21.147946118Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4266814" Oct 13 00:03:21.150974 containerd[1544]: time="2025-10-13T00:03:21.149572473Z" level=info msg="ImageCreate event name:\"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:21.153014 containerd[1544]: time="2025-10-13T00:03:21.152930274Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:21.154426 containerd[1544]: time="2025-10-13T00:03:21.154345850Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5636015\" in 1.714058014s" Oct 13 00:03:21.154426 containerd[1544]: time="2025-10-13T00:03:21.154421017Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\"" Oct 13 00:03:21.161665 containerd[1544]: time="2025-10-13T00:03:21.161580781Z" level=info msg="CreateContainer within sandbox \"80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Oct 13 00:03:21.174728 containerd[1544]: time="2025-10-13T00:03:21.171819200Z" level=info msg="Container 89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:21.186436 containerd[1544]: time="2025-10-13T00:03:21.186035920Z" level=info msg="CreateContainer within sandbox \"80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb\"" Oct 13 00:03:21.188622 containerd[1544]: time="2025-10-13T00:03:21.188560801Z" level=info msg="StartContainer for \"89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb\"" Oct 13 00:03:21.192306 containerd[1544]: time="2025-10-13T00:03:21.192239273Z" level=info msg="connecting to shim 89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb" address="unix:///run/containerd/s/e38ca4d4b5fc773c56eca9a14c886710d14e02ca1e22b19e832656fe9232de0b" protocol=ttrpc version=3 Oct 13 00:03:21.219755 systemd[1]: Started cri-containerd-89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb.scope - libcontainer container 89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb. Oct 13 00:03:21.273461 containerd[1544]: time="2025-10-13T00:03:21.273341467Z" level=info msg="StartContainer for \"89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb\" returns successfully" Oct 13 00:03:21.296152 systemd[1]: cri-containerd-89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb.scope: Deactivated successfully. Oct 13 00:03:21.302592 containerd[1544]: time="2025-10-13T00:03:21.302362482Z" level=info msg="received exit event container_id:\"89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb\" id:\"89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb\" pid:3361 exited_at:{seconds:1760313801 nanos:301691778}" Oct 13 00:03:21.302592 containerd[1544]: time="2025-10-13T00:03:21.302559581Z" level=info msg="TaskExit event in podsandbox handler container_id:\"89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb\" id:\"89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb\" pid:3361 exited_at:{seconds:1760313801 nanos:301691778}" Oct 13 00:03:21.329278 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-89a18805dff2b9614eaa47b6ca4cd10cd99996bf8c3b6377c8c66ca02f3f15fb-rootfs.mount: Deactivated successfully. Oct 13 00:03:21.392436 kubelet[2740]: I1013 00:03:21.391872 2740 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 00:03:21.425185 kubelet[2740]: I1013 00:03:21.424972 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5cbf8867fd-7ll5n" podStartSLOduration=2.674190334 podStartE2EDuration="5.424936883s" podCreationTimestamp="2025-10-13 00:03:16 +0000 UTC" firstStartedPulling="2025-10-13 00:03:16.689149207 +0000 UTC m=+23.616012306" lastFinishedPulling="2025-10-13 00:03:19.439895716 +0000 UTC m=+26.366758855" observedRunningTime="2025-10-13 00:03:20.403373371 +0000 UTC m=+27.330236470" watchObservedRunningTime="2025-10-13 00:03:21.424936883 +0000 UTC m=+28.351799982" Oct 13 00:03:22.239447 kubelet[2740]: E1013 00:03:22.238836 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mtmkk" podUID="c82ba31d-2f24-4bf3-a00d-09a349524d1c" Oct 13 00:03:22.400325 containerd[1544]: time="2025-10-13T00:03:22.400284983Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Oct 13 00:03:24.238922 kubelet[2740]: E1013 00:03:24.238782 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mtmkk" podUID="c82ba31d-2f24-4bf3-a00d-09a349524d1c" Oct 13 00:03:25.826930 containerd[1544]: time="2025-10-13T00:03:25.826868753Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:25.828899 containerd[1544]: time="2025-10-13T00:03:25.828828040Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=65913477" Oct 13 00:03:25.832975 containerd[1544]: time="2025-10-13T00:03:25.830992985Z" level=info msg="ImageCreate event name:\"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:25.834970 containerd[1544]: time="2025-10-13T00:03:25.834926321Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:25.837541 containerd[1544]: time="2025-10-13T00:03:25.837474498Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"67282718\" in 3.436671187s" Oct 13 00:03:25.837791 containerd[1544]: time="2025-10-13T00:03:25.837735200Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\"" Oct 13 00:03:25.841845 containerd[1544]: time="2025-10-13T00:03:25.841798907Z" level=info msg="CreateContainer within sandbox \"80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Oct 13 00:03:25.857540 containerd[1544]: time="2025-10-13T00:03:25.857473285Z" level=info msg="Container bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:25.861530 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2754114025.mount: Deactivated successfully. Oct 13 00:03:25.873796 containerd[1544]: time="2025-10-13T00:03:25.873678349Z" level=info msg="CreateContainer within sandbox \"80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091\"" Oct 13 00:03:25.875135 containerd[1544]: time="2025-10-13T00:03:25.875080348Z" level=info msg="StartContainer for \"bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091\"" Oct 13 00:03:25.879537 containerd[1544]: time="2025-10-13T00:03:25.879485964Z" level=info msg="connecting to shim bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091" address="unix:///run/containerd/s/e38ca4d4b5fc773c56eca9a14c886710d14e02ca1e22b19e832656fe9232de0b" protocol=ttrpc version=3 Oct 13 00:03:25.905629 systemd[1]: Started cri-containerd-bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091.scope - libcontainer container bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091. Oct 13 00:03:25.965041 containerd[1544]: time="2025-10-13T00:03:25.964933779Z" level=info msg="StartContainer for \"bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091\" returns successfully" Oct 13 00:03:26.238728 kubelet[2740]: E1013 00:03:26.238558 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mtmkk" podUID="c82ba31d-2f24-4bf3-a00d-09a349524d1c" Oct 13 00:03:26.513451 containerd[1544]: time="2025-10-13T00:03:26.512855079Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 13 00:03:26.518222 systemd[1]: cri-containerd-bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091.scope: Deactivated successfully. Oct 13 00:03:26.518770 systemd[1]: cri-containerd-bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091.scope: Consumed 557ms CPU time, 187M memory peak, 165.8M written to disk. Oct 13 00:03:26.521450 containerd[1544]: time="2025-10-13T00:03:26.521265618Z" level=info msg="received exit event container_id:\"bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091\" id:\"bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091\" pid:3424 exited_at:{seconds:1760313806 nanos:521012477}" Oct 13 00:03:26.522150 containerd[1544]: time="2025-10-13T00:03:26.522114449Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091\" id:\"bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091\" pid:3424 exited_at:{seconds:1760313806 nanos:521012477}" Oct 13 00:03:26.549905 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bd482b5bc67e57d99e5a161619c0575a56116379b6ff176b1d5590550dbf7091-rootfs.mount: Deactivated successfully. Oct 13 00:03:26.573978 kubelet[2740]: I1013 00:03:26.573921 2740 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Oct 13 00:03:26.638171 kubelet[2740]: I1013 00:03:26.637188 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-ca-bundle\") pod \"whisker-5f864f8bc5-pvdzv\" (UID: \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\") " pod="calico-system/whisker-5f864f8bc5-pvdzv" Oct 13 00:03:26.639340 kubelet[2740]: I1013 00:03:26.638737 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8lq2\" (UniqueName: \"kubernetes.io/projected/ffc90547-d598-495a-8cec-4eb5a2fc13c9-kube-api-access-s8lq2\") pod \"whisker-5f864f8bc5-pvdzv\" (UID: \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\") " pod="calico-system/whisker-5f864f8bc5-pvdzv" Oct 13 00:03:26.639933 kubelet[2740]: I1013 00:03:26.639631 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ad67c87-3ab9-47d8-91ac-3b7ac8198689-config-volume\") pod \"coredns-668d6bf9bc-hls85\" (UID: \"1ad67c87-3ab9-47d8-91ac-3b7ac8198689\") " pod="kube-system/coredns-668d6bf9bc-hls85" Oct 13 00:03:26.641358 kubelet[2740]: I1013 00:03:26.639660 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-backend-key-pair\") pod \"whisker-5f864f8bc5-pvdzv\" (UID: \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\") " pod="calico-system/whisker-5f864f8bc5-pvdzv" Oct 13 00:03:26.643374 kubelet[2740]: I1013 00:03:26.641546 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7v52\" (UniqueName: \"kubernetes.io/projected/1ad67c87-3ab9-47d8-91ac-3b7ac8198689-kube-api-access-v7v52\") pod \"coredns-668d6bf9bc-hls85\" (UID: \"1ad67c87-3ab9-47d8-91ac-3b7ac8198689\") " pod="kube-system/coredns-668d6bf9bc-hls85" Oct 13 00:03:26.643165 systemd[1]: Created slice kubepods-burstable-pod1ad67c87_3ab9_47d8_91ac_3b7ac8198689.slice - libcontainer container kubepods-burstable-pod1ad67c87_3ab9_47d8_91ac_3b7ac8198689.slice. Oct 13 00:03:26.662797 systemd[1]: Created slice kubepods-besteffort-podffc90547_d598_495a_8cec_4eb5a2fc13c9.slice - libcontainer container kubepods-besteffort-podffc90547_d598_495a_8cec_4eb5a2fc13c9.slice. Oct 13 00:03:26.679298 systemd[1]: Created slice kubepods-besteffort-pod675856ae_2bc1_4d35_acbd_777d06e25a59.slice - libcontainer container kubepods-besteffort-pod675856ae_2bc1_4d35_acbd_777d06e25a59.slice. Oct 13 00:03:26.690117 systemd[1]: Created slice kubepods-burstable-podcaf89f26_46b1_4e1a_9e57_9e21ff840fc3.slice - libcontainer container kubepods-burstable-podcaf89f26_46b1_4e1a_9e57_9e21ff840fc3.slice. Oct 13 00:03:26.700738 systemd[1]: Created slice kubepods-besteffort-pod6a443b6d_f54f_4393_a27f_1337e1a343c9.slice - libcontainer container kubepods-besteffort-pod6a443b6d_f54f_4393_a27f_1337e1a343c9.slice. Oct 13 00:03:26.712672 systemd[1]: Created slice kubepods-besteffort-pode9e6f72d_27e8_4bcc_ae87_2bfe7c24625f.slice - libcontainer container kubepods-besteffort-pode9e6f72d_27e8_4bcc_ae87_2bfe7c24625f.slice. Oct 13 00:03:26.723075 systemd[1]: Created slice kubepods-besteffort-podcd2688c6_7370_41d4_a4a6_1d1ea0156d7e.slice - libcontainer container kubepods-besteffort-podcd2688c6_7370_41d4_a4a6_1d1ea0156d7e.slice. Oct 13 00:03:26.745336 kubelet[2740]: I1013 00:03:26.745290 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f-calico-apiserver-certs\") pod \"calico-apiserver-86dfc84df9-nsjp8\" (UID: \"e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f\") " pod="calico-apiserver/calico-apiserver-86dfc84df9-nsjp8" Oct 13 00:03:26.745336 kubelet[2740]: I1013 00:03:26.745352 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a443b6d-f54f-4393-a27f-1337e1a343c9-config\") pod \"goldmane-54d579b49d-8cklf\" (UID: \"6a443b6d-f54f-4393-a27f-1337e1a343c9\") " pod="calico-system/goldmane-54d579b49d-8cklf" Oct 13 00:03:26.746890 kubelet[2740]: I1013 00:03:26.745371 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6vb9\" (UniqueName: \"kubernetes.io/projected/cd2688c6-7370-41d4-a4a6-1d1ea0156d7e-kube-api-access-k6vb9\") pod \"calico-apiserver-86dfc84df9-jd8l6\" (UID: \"cd2688c6-7370-41d4-a4a6-1d1ea0156d7e\") " pod="calico-apiserver/calico-apiserver-86dfc84df9-jd8l6" Oct 13 00:03:26.746890 kubelet[2740]: I1013 00:03:26.745390 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/675856ae-2bc1-4d35-acbd-777d06e25a59-tigera-ca-bundle\") pod \"calico-kube-controllers-76b7656b99-t5tjj\" (UID: \"675856ae-2bc1-4d35-acbd-777d06e25a59\") " pod="calico-system/calico-kube-controllers-76b7656b99-t5tjj" Oct 13 00:03:26.746890 kubelet[2740]: I1013 00:03:26.745436 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwbkr\" (UniqueName: \"kubernetes.io/projected/e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f-kube-api-access-hwbkr\") pod \"calico-apiserver-86dfc84df9-nsjp8\" (UID: \"e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f\") " pod="calico-apiserver/calico-apiserver-86dfc84df9-nsjp8" Oct 13 00:03:26.746890 kubelet[2740]: I1013 00:03:26.745468 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/caf89f26-46b1-4e1a-9e57-9e21ff840fc3-config-volume\") pod \"coredns-668d6bf9bc-v4zvg\" (UID: \"caf89f26-46b1-4e1a-9e57-9e21ff840fc3\") " pod="kube-system/coredns-668d6bf9bc-v4zvg" Oct 13 00:03:26.746890 kubelet[2740]: I1013 00:03:26.745497 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhvrp\" (UniqueName: \"kubernetes.io/projected/caf89f26-46b1-4e1a-9e57-9e21ff840fc3-kube-api-access-lhvrp\") pod \"coredns-668d6bf9bc-v4zvg\" (UID: \"caf89f26-46b1-4e1a-9e57-9e21ff840fc3\") " pod="kube-system/coredns-668d6bf9bc-v4zvg" Oct 13 00:03:26.747038 kubelet[2740]: I1013 00:03:26.745526 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6a443b6d-f54f-4393-a27f-1337e1a343c9-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-8cklf\" (UID: \"6a443b6d-f54f-4393-a27f-1337e1a343c9\") " pod="calico-system/goldmane-54d579b49d-8cklf" Oct 13 00:03:26.747038 kubelet[2740]: I1013 00:03:26.745544 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/6a443b6d-f54f-4393-a27f-1337e1a343c9-goldmane-key-pair\") pod \"goldmane-54d579b49d-8cklf\" (UID: \"6a443b6d-f54f-4393-a27f-1337e1a343c9\") " pod="calico-system/goldmane-54d579b49d-8cklf" Oct 13 00:03:26.747038 kubelet[2740]: I1013 00:03:26.745562 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xdcc\" (UniqueName: \"kubernetes.io/projected/6a443b6d-f54f-4393-a27f-1337e1a343c9-kube-api-access-5xdcc\") pod \"goldmane-54d579b49d-8cklf\" (UID: \"6a443b6d-f54f-4393-a27f-1337e1a343c9\") " pod="calico-system/goldmane-54d579b49d-8cklf" Oct 13 00:03:26.747038 kubelet[2740]: I1013 00:03:26.745579 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kshw\" (UniqueName: \"kubernetes.io/projected/675856ae-2bc1-4d35-acbd-777d06e25a59-kube-api-access-2kshw\") pod \"calico-kube-controllers-76b7656b99-t5tjj\" (UID: \"675856ae-2bc1-4d35-acbd-777d06e25a59\") " pod="calico-system/calico-kube-controllers-76b7656b99-t5tjj" Oct 13 00:03:26.747038 kubelet[2740]: I1013 00:03:26.745601 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/cd2688c6-7370-41d4-a4a6-1d1ea0156d7e-calico-apiserver-certs\") pod \"calico-apiserver-86dfc84df9-jd8l6\" (UID: \"cd2688c6-7370-41d4-a4a6-1d1ea0156d7e\") " pod="calico-apiserver/calico-apiserver-86dfc84df9-jd8l6" Oct 13 00:03:26.959062 containerd[1544]: time="2025-10-13T00:03:26.958693847Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hls85,Uid:1ad67c87-3ab9-47d8-91ac-3b7ac8198689,Namespace:kube-system,Attempt:0,}" Oct 13 00:03:26.970069 containerd[1544]: time="2025-10-13T00:03:26.969481665Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5f864f8bc5-pvdzv,Uid:ffc90547-d598-495a-8cec-4eb5a2fc13c9,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:26.987152 containerd[1544]: time="2025-10-13T00:03:26.987115292Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-76b7656b99-t5tjj,Uid:675856ae-2bc1-4d35-acbd-777d06e25a59,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:26.998698 containerd[1544]: time="2025-10-13T00:03:26.998651971Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-v4zvg,Uid:caf89f26-46b1-4e1a-9e57-9e21ff840fc3,Namespace:kube-system,Attempt:0,}" Oct 13 00:03:27.010982 containerd[1544]: time="2025-10-13T00:03:27.010924933Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-8cklf,Uid:6a443b6d-f54f-4393-a27f-1337e1a343c9,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:27.019490 containerd[1544]: time="2025-10-13T00:03:27.019443984Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-nsjp8,Uid:e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f,Namespace:calico-apiserver,Attempt:0,}" Oct 13 00:03:27.029828 containerd[1544]: time="2025-10-13T00:03:27.029784663Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-jd8l6,Uid:cd2688c6-7370-41d4-a4a6-1d1ea0156d7e,Namespace:calico-apiserver,Attempt:0,}" Oct 13 00:03:27.192742 containerd[1544]: time="2025-10-13T00:03:27.192687403Z" level=error msg="Failed to destroy network for sandbox \"47092645d09dfaeab61139632cee5734168719a6f2d866d8977217c543d5422d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.197365 containerd[1544]: time="2025-10-13T00:03:27.197305777Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5f864f8bc5-pvdzv,Uid:ffc90547-d598-495a-8cec-4eb5a2fc13c9,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"47092645d09dfaeab61139632cee5734168719a6f2d866d8977217c543d5422d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.197910 kubelet[2740]: E1013 00:03:27.197798 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47092645d09dfaeab61139632cee5734168719a6f2d866d8977217c543d5422d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.199296 kubelet[2740]: E1013 00:03:27.197944 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47092645d09dfaeab61139632cee5734168719a6f2d866d8977217c543d5422d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5f864f8bc5-pvdzv" Oct 13 00:03:27.199296 kubelet[2740]: E1013 00:03:27.197968 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47092645d09dfaeab61139632cee5734168719a6f2d866d8977217c543d5422d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5f864f8bc5-pvdzv" Oct 13 00:03:27.199296 kubelet[2740]: E1013 00:03:27.198019 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-5f864f8bc5-pvdzv_calico-system(ffc90547-d598-495a-8cec-4eb5a2fc13c9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-5f864f8bc5-pvdzv_calico-system(ffc90547-d598-495a-8cec-4eb5a2fc13c9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"47092645d09dfaeab61139632cee5734168719a6f2d866d8977217c543d5422d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5f864f8bc5-pvdzv" podUID="ffc90547-d598-495a-8cec-4eb5a2fc13c9" Oct 13 00:03:27.201141 containerd[1544]: time="2025-10-13T00:03:27.200822983Z" level=error msg="Failed to destroy network for sandbox \"42643a65310fe81705b03e48f43be81bfa94b7406317590b5e0ff235b03a98d9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.206111 containerd[1544]: time="2025-10-13T00:03:27.205965080Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-76b7656b99-t5tjj,Uid:675856ae-2bc1-4d35-acbd-777d06e25a59,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"42643a65310fe81705b03e48f43be81bfa94b7406317590b5e0ff235b03a98d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.206829 kubelet[2740]: E1013 00:03:27.206213 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"42643a65310fe81705b03e48f43be81bfa94b7406317590b5e0ff235b03a98d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.206829 kubelet[2740]: E1013 00:03:27.206279 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"42643a65310fe81705b03e48f43be81bfa94b7406317590b5e0ff235b03a98d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-76b7656b99-t5tjj" Oct 13 00:03:27.206829 kubelet[2740]: E1013 00:03:27.206299 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"42643a65310fe81705b03e48f43be81bfa94b7406317590b5e0ff235b03a98d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-76b7656b99-t5tjj" Oct 13 00:03:27.206952 kubelet[2740]: E1013 00:03:27.206348 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-76b7656b99-t5tjj_calico-system(675856ae-2bc1-4d35-acbd-777d06e25a59)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-76b7656b99-t5tjj_calico-system(675856ae-2bc1-4d35-acbd-777d06e25a59)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"42643a65310fe81705b03e48f43be81bfa94b7406317590b5e0ff235b03a98d9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-76b7656b99-t5tjj" podUID="675856ae-2bc1-4d35-acbd-777d06e25a59" Oct 13 00:03:27.218919 containerd[1544]: time="2025-10-13T00:03:27.218310202Z" level=error msg="Failed to destroy network for sandbox \"46a694fc909b2a042b895a79674e73995e1399cfe34d2283e30677d1fbb41972\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.220568 containerd[1544]: time="2025-10-13T00:03:27.220515701Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hls85,Uid:1ad67c87-3ab9-47d8-91ac-3b7ac8198689,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"46a694fc909b2a042b895a79674e73995e1399cfe34d2283e30677d1fbb41972\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.220800 kubelet[2740]: E1013 00:03:27.220747 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"46a694fc909b2a042b895a79674e73995e1399cfe34d2283e30677d1fbb41972\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.220887 kubelet[2740]: E1013 00:03:27.220812 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"46a694fc909b2a042b895a79674e73995e1399cfe34d2283e30677d1fbb41972\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-hls85" Oct 13 00:03:27.220887 kubelet[2740]: E1013 00:03:27.220832 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"46a694fc909b2a042b895a79674e73995e1399cfe34d2283e30677d1fbb41972\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-hls85" Oct 13 00:03:27.220938 kubelet[2740]: E1013 00:03:27.220887 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-hls85_kube-system(1ad67c87-3ab9-47d8-91ac-3b7ac8198689)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-hls85_kube-system(1ad67c87-3ab9-47d8-91ac-3b7ac8198689)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"46a694fc909b2a042b895a79674e73995e1399cfe34d2283e30677d1fbb41972\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-hls85" podUID="1ad67c87-3ab9-47d8-91ac-3b7ac8198689" Oct 13 00:03:27.239036 containerd[1544]: time="2025-10-13T00:03:27.237803184Z" level=error msg="Failed to destroy network for sandbox \"afcbb728ebb292d8fac3bf5ea0a2a17a3de06a74a7c4f55d8c406b613d0117b4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.247200 containerd[1544]: time="2025-10-13T00:03:27.247130101Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-nsjp8,Uid:e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"afcbb728ebb292d8fac3bf5ea0a2a17a3de06a74a7c4f55d8c406b613d0117b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.251502 kubelet[2740]: E1013 00:03:27.250771 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"afcbb728ebb292d8fac3bf5ea0a2a17a3de06a74a7c4f55d8c406b613d0117b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.251502 kubelet[2740]: E1013 00:03:27.250842 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"afcbb728ebb292d8fac3bf5ea0a2a17a3de06a74a7c4f55d8c406b613d0117b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86dfc84df9-nsjp8" Oct 13 00:03:27.251502 kubelet[2740]: E1013 00:03:27.250878 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"afcbb728ebb292d8fac3bf5ea0a2a17a3de06a74a7c4f55d8c406b613d0117b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86dfc84df9-nsjp8" Oct 13 00:03:27.252116 kubelet[2740]: E1013 00:03:27.250934 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86dfc84df9-nsjp8_calico-apiserver(e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86dfc84df9-nsjp8_calico-apiserver(e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"afcbb728ebb292d8fac3bf5ea0a2a17a3de06a74a7c4f55d8c406b613d0117b4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86dfc84df9-nsjp8" podUID="e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f" Oct 13 00:03:27.258974 containerd[1544]: time="2025-10-13T00:03:27.258770725Z" level=error msg="Failed to destroy network for sandbox \"38be903e9fba1f12f69e22d51333c47453d4a0806d37930381c883145dead7b8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.262169 containerd[1544]: time="2025-10-13T00:03:27.262020189Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-v4zvg,Uid:caf89f26-46b1-4e1a-9e57-9e21ff840fc3,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"38be903e9fba1f12f69e22d51333c47453d4a0806d37930381c883145dead7b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.262839 kubelet[2740]: E1013 00:03:27.262763 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38be903e9fba1f12f69e22d51333c47453d4a0806d37930381c883145dead7b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.262839 kubelet[2740]: E1013 00:03:27.262825 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38be903e9fba1f12f69e22d51333c47453d4a0806d37930381c883145dead7b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-v4zvg" Oct 13 00:03:27.262839 kubelet[2740]: E1013 00:03:27.262897 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38be903e9fba1f12f69e22d51333c47453d4a0806d37930381c883145dead7b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-v4zvg" Oct 13 00:03:27.264113 kubelet[2740]: E1013 00:03:27.262949 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-v4zvg_kube-system(caf89f26-46b1-4e1a-9e57-9e21ff840fc3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-v4zvg_kube-system(caf89f26-46b1-4e1a-9e57-9e21ff840fc3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"38be903e9fba1f12f69e22d51333c47453d4a0806d37930381c883145dead7b8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-v4zvg" podUID="caf89f26-46b1-4e1a-9e57-9e21ff840fc3" Oct 13 00:03:27.269785 containerd[1544]: time="2025-10-13T00:03:27.269373346Z" level=error msg="Failed to destroy network for sandbox \"5e784a40d94082fdcba642d12f2ab9784d967417192bc8050343f1fbee2906ce\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.272924 containerd[1544]: time="2025-10-13T00:03:27.272579606Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-jd8l6,Uid:cd2688c6-7370-41d4-a4a6-1d1ea0156d7e,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5e784a40d94082fdcba642d12f2ab9784d967417192bc8050343f1fbee2906ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.273509 kubelet[2740]: E1013 00:03:27.273436 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5e784a40d94082fdcba642d12f2ab9784d967417192bc8050343f1fbee2906ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.273509 kubelet[2740]: E1013 00:03:27.273498 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5e784a40d94082fdcba642d12f2ab9784d967417192bc8050343f1fbee2906ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86dfc84df9-jd8l6" Oct 13 00:03:27.273509 kubelet[2740]: E1013 00:03:27.273534 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5e784a40d94082fdcba642d12f2ab9784d967417192bc8050343f1fbee2906ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86dfc84df9-jd8l6" Oct 13 00:03:27.274120 kubelet[2740]: E1013 00:03:27.273584 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86dfc84df9-jd8l6_calico-apiserver(cd2688c6-7370-41d4-a4a6-1d1ea0156d7e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86dfc84df9-jd8l6_calico-apiserver(cd2688c6-7370-41d4-a4a6-1d1ea0156d7e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5e784a40d94082fdcba642d12f2ab9784d967417192bc8050343f1fbee2906ce\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86dfc84df9-jd8l6" podUID="cd2688c6-7370-41d4-a4a6-1d1ea0156d7e" Oct 13 00:03:27.287544 containerd[1544]: time="2025-10-13T00:03:27.286629826Z" level=error msg="Failed to destroy network for sandbox \"16c0ee98c742189ba0786e70065de24e0c0cbe51bf581a326adb9b9741244cdc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.288899 containerd[1544]: time="2025-10-13T00:03:27.288815803Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-8cklf,Uid:6a443b6d-f54f-4393-a27f-1337e1a343c9,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"16c0ee98c742189ba0786e70065de24e0c0cbe51bf581a326adb9b9741244cdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.289393 kubelet[2740]: E1013 00:03:27.289357 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"16c0ee98c742189ba0786e70065de24e0c0cbe51bf581a326adb9b9741244cdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:27.289678 kubelet[2740]: E1013 00:03:27.289509 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"16c0ee98c742189ba0786e70065de24e0c0cbe51bf581a326adb9b9741244cdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-8cklf" Oct 13 00:03:27.289678 kubelet[2740]: E1013 00:03:27.289533 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"16c0ee98c742189ba0786e70065de24e0c0cbe51bf581a326adb9b9741244cdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-8cklf" Oct 13 00:03:27.289678 kubelet[2740]: E1013 00:03:27.289582 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-8cklf_calico-system(6a443b6d-f54f-4393-a27f-1337e1a343c9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-8cklf_calico-system(6a443b6d-f54f-4393-a27f-1337e1a343c9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"16c0ee98c742189ba0786e70065de24e0c0cbe51bf581a326adb9b9741244cdc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-8cklf" podUID="6a443b6d-f54f-4393-a27f-1337e1a343c9" Oct 13 00:03:27.427997 containerd[1544]: time="2025-10-13T00:03:27.427833044Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Oct 13 00:03:28.254307 systemd[1]: Created slice kubepods-besteffort-podc82ba31d_2f24_4bf3_a00d_09a349524d1c.slice - libcontainer container kubepods-besteffort-podc82ba31d_2f24_4bf3_a00d_09a349524d1c.slice. Oct 13 00:03:28.259501 containerd[1544]: time="2025-10-13T00:03:28.259108494Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mtmkk,Uid:c82ba31d-2f24-4bf3-a00d-09a349524d1c,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:28.329449 containerd[1544]: time="2025-10-13T00:03:28.329283135Z" level=error msg="Failed to destroy network for sandbox \"9145e4e9c1b7d445846c19a7a70dd2097400eddec1493d6a4b1f1b9baeadd0cf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:28.333298 containerd[1544]: time="2025-10-13T00:03:28.333057994Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mtmkk,Uid:c82ba31d-2f24-4bf3-a00d-09a349524d1c,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9145e4e9c1b7d445846c19a7a70dd2097400eddec1493d6a4b1f1b9baeadd0cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:28.334270 kubelet[2740]: E1013 00:03:28.333452 2740 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9145e4e9c1b7d445846c19a7a70dd2097400eddec1493d6a4b1f1b9baeadd0cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 00:03:28.334270 kubelet[2740]: E1013 00:03:28.333698 2740 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9145e4e9c1b7d445846c19a7a70dd2097400eddec1493d6a4b1f1b9baeadd0cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-mtmkk" Oct 13 00:03:28.334270 kubelet[2740]: E1013 00:03:28.333727 2740 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9145e4e9c1b7d445846c19a7a70dd2097400eddec1493d6a4b1f1b9baeadd0cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-mtmkk" Oct 13 00:03:28.336601 kubelet[2740]: E1013 00:03:28.333778 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-mtmkk_calico-system(c82ba31d-2f24-4bf3-a00d-09a349524d1c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-mtmkk_calico-system(c82ba31d-2f24-4bf3-a00d-09a349524d1c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9145e4e9c1b7d445846c19a7a70dd2097400eddec1493d6a4b1f1b9baeadd0cf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-mtmkk" podUID="c82ba31d-2f24-4bf3-a00d-09a349524d1c" Oct 13 00:03:28.334640 systemd[1]: run-netns-cni\x2d16813197\x2d2887\x2d9932\x2d7a8a\x2d8633482fa4d5.mount: Deactivated successfully. Oct 13 00:03:34.394077 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3591164682.mount: Deactivated successfully. Oct 13 00:03:34.419458 containerd[1544]: time="2025-10-13T00:03:34.418903090Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:34.420275 containerd[1544]: time="2025-10-13T00:03:34.420224223Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=151100457" Oct 13 00:03:34.420806 containerd[1544]: time="2025-10-13T00:03:34.420774821Z" level=info msg="ImageCreate event name:\"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:34.423952 containerd[1544]: time="2025-10-13T00:03:34.423858797Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:34.424779 containerd[1544]: time="2025-10-13T00:03:34.424390194Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"151100319\" in 6.996131876s" Oct 13 00:03:34.424779 containerd[1544]: time="2025-10-13T00:03:34.424452879Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\"" Oct 13 00:03:34.447035 containerd[1544]: time="2025-10-13T00:03:34.446990337Z" level=info msg="CreateContainer within sandbox \"80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Oct 13 00:03:34.467431 containerd[1544]: time="2025-10-13T00:03:34.465617601Z" level=info msg="Container 2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:34.478977 containerd[1544]: time="2025-10-13T00:03:34.478913812Z" level=info msg="CreateContainer within sandbox \"80f7b35c05df2a57d3d3eb7836394167fe4c3e7b3e5d112b957d65a63c7b6ac9\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\"" Oct 13 00:03:34.481218 containerd[1544]: time="2025-10-13T00:03:34.481183931Z" level=info msg="StartContainer for \"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\"" Oct 13 00:03:34.483537 containerd[1544]: time="2025-10-13T00:03:34.483486693Z" level=info msg="connecting to shim 2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02" address="unix:///run/containerd/s/e38ca4d4b5fc773c56eca9a14c886710d14e02ca1e22b19e832656fe9232de0b" protocol=ttrpc version=3 Oct 13 00:03:34.512692 systemd[1]: Started cri-containerd-2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02.scope - libcontainer container 2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02. Oct 13 00:03:34.564958 containerd[1544]: time="2025-10-13T00:03:34.564575691Z" level=info msg="StartContainer for \"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" returns successfully" Oct 13 00:03:34.733758 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Oct 13 00:03:34.733927 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Oct 13 00:03:35.014030 kubelet[2740]: I1013 00:03:35.013563 2740 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-ca-bundle\") pod \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\" (UID: \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\") " Oct 13 00:03:35.015088 kubelet[2740]: I1013 00:03:35.014614 2740 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8lq2\" (UniqueName: \"kubernetes.io/projected/ffc90547-d598-495a-8cec-4eb5a2fc13c9-kube-api-access-s8lq2\") pod \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\" (UID: \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\") " Oct 13 00:03:35.015088 kubelet[2740]: I1013 00:03:35.014651 2740 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-backend-key-pair\") pod \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\" (UID: \"ffc90547-d598-495a-8cec-4eb5a2fc13c9\") " Oct 13 00:03:35.015088 kubelet[2740]: I1013 00:03:35.013980 2740 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "ffc90547-d598-495a-8cec-4eb5a2fc13c9" (UID: "ffc90547-d598-495a-8cec-4eb5a2fc13c9"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Oct 13 00:03:35.020957 kubelet[2740]: I1013 00:03:35.020915 2740 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc90547-d598-495a-8cec-4eb5a2fc13c9-kube-api-access-s8lq2" (OuterVolumeSpecName: "kube-api-access-s8lq2") pod "ffc90547-d598-495a-8cec-4eb5a2fc13c9" (UID: "ffc90547-d598-495a-8cec-4eb5a2fc13c9"). InnerVolumeSpecName "kube-api-access-s8lq2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Oct 13 00:03:35.021680 kubelet[2740]: I1013 00:03:35.021640 2740 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "ffc90547-d598-495a-8cec-4eb5a2fc13c9" (UID: "ffc90547-d598-495a-8cec-4eb5a2fc13c9"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Oct 13 00:03:35.115790 kubelet[2740]: I1013 00:03:35.115747 2740 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-s8lq2\" (UniqueName: \"kubernetes.io/projected/ffc90547-d598-495a-8cec-4eb5a2fc13c9-kube-api-access-s8lq2\") on node \"ci-4459-1-0-3-bea61e6062\" DevicePath \"\"" Oct 13 00:03:35.115790 kubelet[2740]: I1013 00:03:35.115783 2740 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-backend-key-pair\") on node \"ci-4459-1-0-3-bea61e6062\" DevicePath \"\"" Oct 13 00:03:35.115790 kubelet[2740]: I1013 00:03:35.115794 2740 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffc90547-d598-495a-8cec-4eb5a2fc13c9-whisker-ca-bundle\") on node \"ci-4459-1-0-3-bea61e6062\" DevicePath \"\"" Oct 13 00:03:35.250003 systemd[1]: Removed slice kubepods-besteffort-podffc90547_d598_495a_8cec_4eb5a2fc13c9.slice - libcontainer container kubepods-besteffort-podffc90547_d598_495a_8cec_4eb5a2fc13c9.slice. Oct 13 00:03:35.395292 systemd[1]: var-lib-kubelet-pods-ffc90547\x2dd598\x2d495a\x2d8cec\x2d4eb5a2fc13c9-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ds8lq2.mount: Deactivated successfully. Oct 13 00:03:35.395386 systemd[1]: var-lib-kubelet-pods-ffc90547\x2dd598\x2d495a\x2d8cec\x2d4eb5a2fc13c9-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Oct 13 00:03:35.511844 kubelet[2740]: I1013 00:03:35.511515 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-hkq4m" podStartSLOduration=1.876323585 podStartE2EDuration="19.51134189s" podCreationTimestamp="2025-10-13 00:03:16 +0000 UTC" firstStartedPulling="2025-10-13 00:03:16.790757346 +0000 UTC m=+23.717620445" lastFinishedPulling="2025-10-13 00:03:34.425775691 +0000 UTC m=+41.352638750" observedRunningTime="2025-10-13 00:03:35.507863011 +0000 UTC m=+42.434726110" watchObservedRunningTime="2025-10-13 00:03:35.51134189 +0000 UTC m=+42.438205029" Oct 13 00:03:35.606528 systemd[1]: Created slice kubepods-besteffort-podb35791e5_da21_4fa2_bab0_692d0a1f1274.slice - libcontainer container kubepods-besteffort-podb35791e5_da21_4fa2_bab0_692d0a1f1274.slice. Oct 13 00:03:35.619960 kubelet[2740]: I1013 00:03:35.619863 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/b35791e5-da21-4fa2-bab0-692d0a1f1274-whisker-backend-key-pair\") pod \"whisker-57df5ddf6b-bw8z2\" (UID: \"b35791e5-da21-4fa2-bab0-692d0a1f1274\") " pod="calico-system/whisker-57df5ddf6b-bw8z2" Oct 13 00:03:35.620609 kubelet[2740]: I1013 00:03:35.620005 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b35791e5-da21-4fa2-bab0-692d0a1f1274-whisker-ca-bundle\") pod \"whisker-57df5ddf6b-bw8z2\" (UID: \"b35791e5-da21-4fa2-bab0-692d0a1f1274\") " pod="calico-system/whisker-57df5ddf6b-bw8z2" Oct 13 00:03:35.620609 kubelet[2740]: I1013 00:03:35.620034 2740 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sbb6\" (UniqueName: \"kubernetes.io/projected/b35791e5-da21-4fa2-bab0-692d0a1f1274-kube-api-access-8sbb6\") pod \"whisker-57df5ddf6b-bw8z2\" (UID: \"b35791e5-da21-4fa2-bab0-692d0a1f1274\") " pod="calico-system/whisker-57df5ddf6b-bw8z2" Oct 13 00:03:35.700220 containerd[1544]: time="2025-10-13T00:03:35.700006192Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"006064abbf96de96e4fa0a5d3ac86dedc5cafd4533a24e6b90df8bc2fa184f88\" pid:3763 exit_status:1 exited_at:{seconds:1760313815 nanos:699044646}" Oct 13 00:03:35.913908 containerd[1544]: time="2025-10-13T00:03:35.913388875Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-57df5ddf6b-bw8z2,Uid:b35791e5-da21-4fa2-bab0-692d0a1f1274,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:36.174459 systemd-networkd[1418]: cali769ff1768f0: Link UP Oct 13 00:03:36.175791 systemd-networkd[1418]: cali769ff1768f0: Gained carrier Oct 13 00:03:36.207123 containerd[1544]: 2025-10-13 00:03:35.978 [INFO][3777] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 13 00:03:36.207123 containerd[1544]: 2025-10-13 00:03:36.024 [INFO][3777] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0 whisker-57df5ddf6b- calico-system b35791e5-da21-4fa2-bab0-692d0a1f1274 869 0 2025-10-13 00:03:35 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:57df5ddf6b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 whisker-57df5ddf6b-bw8z2 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali769ff1768f0 [] [] }} ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-" Oct 13 00:03:36.207123 containerd[1544]: 2025-10-13 00:03:36.024 [INFO][3777] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" Oct 13 00:03:36.207123 containerd[1544]: 2025-10-13 00:03:36.092 [INFO][3790] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" HandleID="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Workload="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.092 [INFO][3790] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" HandleID="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Workload="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400024aff0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-1-0-3-bea61e6062", "pod":"whisker-57df5ddf6b-bw8z2", "timestamp":"2025-10-13 00:03:36.092716151 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.093 [INFO][3790] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.093 [INFO][3790] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.093 [INFO][3790] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.105 [INFO][3790] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.117 [INFO][3790] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.129 [INFO][3790] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.133 [INFO][3790] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.207377 containerd[1544]: 2025-10-13 00:03:36.137 [INFO][3790] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.208717 containerd[1544]: 2025-10-13 00:03:36.137 [INFO][3790] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.208717 containerd[1544]: 2025-10-13 00:03:36.140 [INFO][3790] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc Oct 13 00:03:36.208717 containerd[1544]: 2025-10-13 00:03:36.147 [INFO][3790] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.208717 containerd[1544]: 2025-10-13 00:03:36.156 [INFO][3790] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.1/26] block=192.168.83.0/26 handle="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.208717 containerd[1544]: 2025-10-13 00:03:36.156 [INFO][3790] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.1/26] handle="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:36.208717 containerd[1544]: 2025-10-13 00:03:36.156 [INFO][3790] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:36.208717 containerd[1544]: 2025-10-13 00:03:36.156 [INFO][3790] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.1/26] IPv6=[] ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" HandleID="k8s-pod-network.307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Workload="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" Oct 13 00:03:36.208895 containerd[1544]: 2025-10-13 00:03:36.162 [INFO][3777] cni-plugin/k8s.go 418: Populated endpoint ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0", GenerateName:"whisker-57df5ddf6b-", Namespace:"calico-system", SelfLink:"", UID:"b35791e5-da21-4fa2-bab0-692d0a1f1274", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"57df5ddf6b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"whisker-57df5ddf6b-bw8z2", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.83.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali769ff1768f0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:36.208895 containerd[1544]: 2025-10-13 00:03:36.163 [INFO][3777] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.1/32] ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" Oct 13 00:03:36.208999 containerd[1544]: 2025-10-13 00:03:36.163 [INFO][3777] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali769ff1768f0 ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" Oct 13 00:03:36.208999 containerd[1544]: 2025-10-13 00:03:36.177 [INFO][3777] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" Oct 13 00:03:36.209041 containerd[1544]: 2025-10-13 00:03:36.178 [INFO][3777] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0", GenerateName:"whisker-57df5ddf6b-", Namespace:"calico-system", SelfLink:"", UID:"b35791e5-da21-4fa2-bab0-692d0a1f1274", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"57df5ddf6b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc", Pod:"whisker-57df5ddf6b-bw8z2", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.83.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali769ff1768f0", MAC:"16:a0:a3:48:72:27", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:36.209092 containerd[1544]: 2025-10-13 00:03:36.204 [INFO][3777] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" Namespace="calico-system" Pod="whisker-57df5ddf6b-bw8z2" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-whisker--57df5ddf6b--bw8z2-eth0" Oct 13 00:03:36.271429 containerd[1544]: time="2025-10-13T00:03:36.270972213Z" level=info msg="connecting to shim 307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc" address="unix:///run/containerd/s/fbf5c6ae55aaf257ba4c10d543a75529b712644b7b1e0040084f20b94120d2a4" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:36.362755 systemd[1]: Started cri-containerd-307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc.scope - libcontainer container 307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc. Oct 13 00:03:36.512287 containerd[1544]: time="2025-10-13T00:03:36.511818831Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-57df5ddf6b-bw8z2,Uid:b35791e5-da21-4fa2-bab0-692d0a1f1274,Namespace:calico-system,Attempt:0,} returns sandbox id \"307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc\"" Oct 13 00:03:36.518874 containerd[1544]: time="2025-10-13T00:03:36.518718618Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Oct 13 00:03:36.747361 containerd[1544]: time="2025-10-13T00:03:36.747201719Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"a9f194dd46144982015affeaf70977d29d91c7e954c4880a47e04030147f6700\" pid:3947 exit_status:1 exited_at:{seconds:1760313816 nanos:746734448}" Oct 13 00:03:37.242685 kubelet[2740]: I1013 00:03:37.242239 2740 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc90547-d598-495a-8cec-4eb5a2fc13c9" path="/var/lib/kubelet/pods/ffc90547-d598-495a-8cec-4eb5a2fc13c9/volumes" Oct 13 00:03:37.534691 systemd-networkd[1418]: cali769ff1768f0: Gained IPv6LL Oct 13 00:03:37.603164 containerd[1544]: time="2025-10-13T00:03:37.603120881Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"eea56026abc8b122c0259ea96e36d5423b0d432457b59f604c4d6ddba47b82f3\" pid:3979 exit_status:1 exited_at:{seconds:1760313817 nanos:602798339}" Oct 13 00:03:38.232134 containerd[1544]: time="2025-10-13T00:03:38.232061897Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:38.233970 containerd[1544]: time="2025-10-13T00:03:38.233912379Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4605606" Oct 13 00:03:38.234791 containerd[1544]: time="2025-10-13T00:03:38.234742593Z" level=info msg="ImageCreate event name:\"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:38.237964 containerd[1544]: time="2025-10-13T00:03:38.237921362Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:38.239583 containerd[1544]: time="2025-10-13T00:03:38.239519426Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"5974839\" in 1.720507869s" Oct 13 00:03:38.239677 containerd[1544]: time="2025-10-13T00:03:38.239581911Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\"" Oct 13 00:03:38.242517 containerd[1544]: time="2025-10-13T00:03:38.242475500Z" level=info msg="CreateContainer within sandbox \"307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Oct 13 00:03:38.254491 containerd[1544]: time="2025-10-13T00:03:38.253589989Z" level=info msg="Container 8690a4e1316554347ee503d859230a3d52f3fd99f5fd5daa4f2d949667e41ab3: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:38.260892 kubelet[2740]: I1013 00:03:38.259848 2740 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 00:03:38.271327 containerd[1544]: time="2025-10-13T00:03:38.271249348Z" level=info msg="CreateContainer within sandbox \"307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"8690a4e1316554347ee503d859230a3d52f3fd99f5fd5daa4f2d949667e41ab3\"" Oct 13 00:03:38.277182 containerd[1544]: time="2025-10-13T00:03:38.277131734Z" level=info msg="StartContainer for \"8690a4e1316554347ee503d859230a3d52f3fd99f5fd5daa4f2d949667e41ab3\"" Oct 13 00:03:38.282490 containerd[1544]: time="2025-10-13T00:03:38.282423561Z" level=info msg="connecting to shim 8690a4e1316554347ee503d859230a3d52f3fd99f5fd5daa4f2d949667e41ab3" address="unix:///run/containerd/s/fbf5c6ae55aaf257ba4c10d543a75529b712644b7b1e0040084f20b94120d2a4" protocol=ttrpc version=3 Oct 13 00:03:38.331715 systemd[1]: Started cri-containerd-8690a4e1316554347ee503d859230a3d52f3fd99f5fd5daa4f2d949667e41ab3.scope - libcontainer container 8690a4e1316554347ee503d859230a3d52f3fd99f5fd5daa4f2d949667e41ab3. Oct 13 00:03:38.397117 containerd[1544]: time="2025-10-13T00:03:38.397072321Z" level=info msg="StartContainer for \"8690a4e1316554347ee503d859230a3d52f3fd99f5fd5daa4f2d949667e41ab3\" returns successfully" Oct 13 00:03:38.401456 containerd[1544]: time="2025-10-13T00:03:38.400588752Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Oct 13 00:03:39.240846 containerd[1544]: time="2025-10-13T00:03:39.239589004Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-76b7656b99-t5tjj,Uid:675856ae-2bc1-4d35-acbd-777d06e25a59,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:39.428885 systemd-networkd[1418]: cali58bfa45b484: Link UP Oct 13 00:03:39.431529 systemd-networkd[1418]: cali58bfa45b484: Gained carrier Oct 13 00:03:39.461147 containerd[1544]: 2025-10-13 00:03:39.312 [INFO][4107] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0 calico-kube-controllers-76b7656b99- calico-system 675856ae-2bc1-4d35-acbd-777d06e25a59 804 0 2025-10-13 00:03:16 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:76b7656b99 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 calico-kube-controllers-76b7656b99-t5tjj eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali58bfa45b484 [] [] }} ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-" Oct 13 00:03:39.461147 containerd[1544]: 2025-10-13 00:03:39.312 [INFO][4107] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" Oct 13 00:03:39.461147 containerd[1544]: 2025-10-13 00:03:39.351 [INFO][4119] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" HandleID="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.352 [INFO][4119] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" HandleID="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400024b0a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-1-0-3-bea61e6062", "pod":"calico-kube-controllers-76b7656b99-t5tjj", "timestamp":"2025-10-13 00:03:39.351880264 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.352 [INFO][4119] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.352 [INFO][4119] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.352 [INFO][4119] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.367 [INFO][4119] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.375 [INFO][4119] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.382 [INFO][4119] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.386 [INFO][4119] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461391 containerd[1544]: 2025-10-13 00:03:39.389 [INFO][4119] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461697 containerd[1544]: 2025-10-13 00:03:39.390 [INFO][4119] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461697 containerd[1544]: 2025-10-13 00:03:39.392 [INFO][4119] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933 Oct 13 00:03:39.461697 containerd[1544]: 2025-10-13 00:03:39.403 [INFO][4119] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461697 containerd[1544]: 2025-10-13 00:03:39.410 [INFO][4119] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.2/26] block=192.168.83.0/26 handle="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461697 containerd[1544]: 2025-10-13 00:03:39.410 [INFO][4119] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.2/26] handle="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:39.461697 containerd[1544]: 2025-10-13 00:03:39.410 [INFO][4119] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:39.461697 containerd[1544]: 2025-10-13 00:03:39.410 [INFO][4119] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.2/26] IPv6=[] ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" HandleID="k8s-pod-network.fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" Oct 13 00:03:39.461904 containerd[1544]: 2025-10-13 00:03:39.415 [INFO][4107] cni-plugin/k8s.go 418: Populated endpoint ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0", GenerateName:"calico-kube-controllers-76b7656b99-", Namespace:"calico-system", SelfLink:"", UID:"675856ae-2bc1-4d35-acbd-777d06e25a59", ResourceVersion:"804", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"76b7656b99", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"calico-kube-controllers-76b7656b99-t5tjj", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.83.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali58bfa45b484", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:39.461966 containerd[1544]: 2025-10-13 00:03:39.421 [INFO][4107] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.2/32] ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" Oct 13 00:03:39.461966 containerd[1544]: 2025-10-13 00:03:39.421 [INFO][4107] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali58bfa45b484 ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" Oct 13 00:03:39.461966 containerd[1544]: 2025-10-13 00:03:39.431 [INFO][4107] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" Oct 13 00:03:39.462050 containerd[1544]: 2025-10-13 00:03:39.432 [INFO][4107] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0", GenerateName:"calico-kube-controllers-76b7656b99-", Namespace:"calico-system", SelfLink:"", UID:"675856ae-2bc1-4d35-acbd-777d06e25a59", ResourceVersion:"804", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"76b7656b99", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933", Pod:"calico-kube-controllers-76b7656b99-t5tjj", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.83.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali58bfa45b484", MAC:"92:a4:ce:2e:46:75", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:39.462172 containerd[1544]: 2025-10-13 00:03:39.454 [INFO][4107] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" Namespace="calico-system" Pod="calico-kube-controllers-76b7656b99-t5tjj" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--kube--controllers--76b7656b99--t5tjj-eth0" Oct 13 00:03:39.507998 containerd[1544]: time="2025-10-13T00:03:39.507271191Z" level=info msg="connecting to shim fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933" address="unix:///run/containerd/s/b8f314fd7661da5a49ebb1cef9d8c6b06146a540f254b3de0bac789f4347ed63" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:39.566206 systemd-networkd[1418]: vxlan.calico: Link UP Oct 13 00:03:39.566215 systemd-networkd[1418]: vxlan.calico: Gained carrier Oct 13 00:03:39.590108 systemd[1]: Started cri-containerd-fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933.scope - libcontainer container fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933. Oct 13 00:03:39.684422 containerd[1544]: time="2025-10-13T00:03:39.682738936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-76b7656b99-t5tjj,Uid:675856ae-2bc1-4d35-acbd-777d06e25a59,Namespace:calico-system,Attempt:0,} returns sandbox id \"fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933\"" Oct 13 00:03:41.241112 containerd[1544]: time="2025-10-13T00:03:41.240828107Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hls85,Uid:1ad67c87-3ab9-47d8-91ac-3b7ac8198689,Namespace:kube-system,Attempt:0,}" Oct 13 00:03:41.242285 containerd[1544]: time="2025-10-13T00:03:41.240832907Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-v4zvg,Uid:caf89f26-46b1-4e1a-9e57-9e21ff840fc3,Namespace:kube-system,Attempt:0,}" Oct 13 00:03:41.243512 containerd[1544]: time="2025-10-13T00:03:41.243462433Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mtmkk,Uid:c82ba31d-2f24-4bf3-a00d-09a349524d1c,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:41.244064 containerd[1544]: time="2025-10-13T00:03:41.244037069Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-nsjp8,Uid:e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f,Namespace:calico-apiserver,Attempt:0,}" Oct 13 00:03:41.244965 containerd[1544]: time="2025-10-13T00:03:41.244655908Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-jd8l6,Uid:cd2688c6-7370-41d4-a4a6-1d1ea0156d7e,Namespace:calico-apiserver,Attempt:0,}" Oct 13 00:03:41.375672 systemd-networkd[1418]: cali58bfa45b484: Gained IPv6LL Oct 13 00:03:41.439720 systemd-networkd[1418]: vxlan.calico: Gained IPv6LL Oct 13 00:03:41.665977 systemd-networkd[1418]: cali20d2d13c1b6: Link UP Oct 13 00:03:41.667328 systemd-networkd[1418]: cali20d2d13c1b6: Gained carrier Oct 13 00:03:41.708375 containerd[1544]: 2025-10-13 00:03:41.446 [INFO][4273] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0 calico-apiserver-86dfc84df9- calico-apiserver cd2688c6-7370-41d4-a4a6-1d1ea0156d7e 805 0 2025-10-13 00:03:11 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:86dfc84df9 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 calico-apiserver-86dfc84df9-jd8l6 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali20d2d13c1b6 [] [] }} ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-" Oct 13 00:03:41.708375 containerd[1544]: 2025-10-13 00:03:41.446 [INFO][4273] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" Oct 13 00:03:41.708375 containerd[1544]: 2025-10-13 00:03:41.532 [INFO][4323] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" HandleID="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.533 [INFO][4323] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" HandleID="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000318410), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4459-1-0-3-bea61e6062", "pod":"calico-apiserver-86dfc84df9-jd8l6", "timestamp":"2025-10-13 00:03:41.531549368 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.533 [INFO][4323] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.533 [INFO][4323] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.533 [INFO][4323] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.566 [INFO][4323] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.580 [INFO][4323] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.590 [INFO][4323] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.601 [INFO][4323] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.708612 containerd[1544]: 2025-10-13 00:03:41.608 [INFO][4323] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.709120 containerd[1544]: 2025-10-13 00:03:41.608 [INFO][4323] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.709120 containerd[1544]: 2025-10-13 00:03:41.611 [INFO][4323] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882 Oct 13 00:03:41.709120 containerd[1544]: 2025-10-13 00:03:41.619 [INFO][4323] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.709120 containerd[1544]: 2025-10-13 00:03:41.635 [INFO][4323] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.3/26] block=192.168.83.0/26 handle="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.709120 containerd[1544]: 2025-10-13 00:03:41.635 [INFO][4323] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.3/26] handle="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.709120 containerd[1544]: 2025-10-13 00:03:41.635 [INFO][4323] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:41.709120 containerd[1544]: 2025-10-13 00:03:41.635 [INFO][4323] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.3/26] IPv6=[] ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" HandleID="k8s-pod-network.0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" Oct 13 00:03:41.709259 containerd[1544]: 2025-10-13 00:03:41.652 [INFO][4273] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0", GenerateName:"calico-apiserver-86dfc84df9-", Namespace:"calico-apiserver", SelfLink:"", UID:"cd2688c6-7370-41d4-a4a6-1d1ea0156d7e", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86dfc84df9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"calico-apiserver-86dfc84df9-jd8l6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.83.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali20d2d13c1b6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:41.709312 containerd[1544]: 2025-10-13 00:03:41.653 [INFO][4273] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.3/32] ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" Oct 13 00:03:41.709312 containerd[1544]: 2025-10-13 00:03:41.653 [INFO][4273] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali20d2d13c1b6 ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" Oct 13 00:03:41.709312 containerd[1544]: 2025-10-13 00:03:41.667 [INFO][4273] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" Oct 13 00:03:41.709435 containerd[1544]: 2025-10-13 00:03:41.677 [INFO][4273] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0", GenerateName:"calico-apiserver-86dfc84df9-", Namespace:"calico-apiserver", SelfLink:"", UID:"cd2688c6-7370-41d4-a4a6-1d1ea0156d7e", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86dfc84df9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882", Pod:"calico-apiserver-86dfc84df9-jd8l6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.83.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali20d2d13c1b6", MAC:"52:9c:74:42:85:b9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:41.709496 containerd[1544]: 2025-10-13 00:03:41.697 [INFO][4273] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-jd8l6" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--jd8l6-eth0" Oct 13 00:03:41.796181 systemd-networkd[1418]: cali9b5fd7bc851: Link UP Oct 13 00:03:41.798518 systemd-networkd[1418]: cali9b5fd7bc851: Gained carrier Oct 13 00:03:41.805386 containerd[1544]: time="2025-10-13T00:03:41.805283959Z" level=info msg="connecting to shim 0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882" address="unix:///run/containerd/s/e17e363542aede78e81f6544e8f4ce021bc5a7065a82552e161522fbedbbef75" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:41.838425 containerd[1544]: 2025-10-13 00:03:41.456 [INFO][4265] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0 calico-apiserver-86dfc84df9- calico-apiserver e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f 802 0 2025-10-13 00:03:11 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:86dfc84df9 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 calico-apiserver-86dfc84df9-nsjp8 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali9b5fd7bc851 [] [] }} ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-" Oct 13 00:03:41.838425 containerd[1544]: 2025-10-13 00:03:41.456 [INFO][4265] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" Oct 13 00:03:41.838425 containerd[1544]: 2025-10-13 00:03:41.562 [INFO][4335] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" HandleID="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.562 [INFO][4335] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" HandleID="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002e7890), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4459-1-0-3-bea61e6062", "pod":"calico-apiserver-86dfc84df9-nsjp8", "timestamp":"2025-10-13 00:03:41.562048287 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.562 [INFO][4335] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.637 [INFO][4335] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.637 [INFO][4335] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.672 [INFO][4335] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.694 [INFO][4335] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.711 [INFO][4335] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.716 [INFO][4335] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.838973 containerd[1544]: 2025-10-13 00:03:41.722 [INFO][4335] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.840688 containerd[1544]: 2025-10-13 00:03:41.722 [INFO][4335] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.840688 containerd[1544]: 2025-10-13 00:03:41.726 [INFO][4335] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659 Oct 13 00:03:41.840688 containerd[1544]: 2025-10-13 00:03:41.737 [INFO][4335] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.840688 containerd[1544]: 2025-10-13 00:03:41.751 [INFO][4335] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.4/26] block=192.168.83.0/26 handle="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.840688 containerd[1544]: 2025-10-13 00:03:41.751 [INFO][4335] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.4/26] handle="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.840688 containerd[1544]: 2025-10-13 00:03:41.752 [INFO][4335] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:41.840688 containerd[1544]: 2025-10-13 00:03:41.752 [INFO][4335] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.4/26] IPv6=[] ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" HandleID="k8s-pod-network.74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Workload="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" Oct 13 00:03:41.840936 containerd[1544]: 2025-10-13 00:03:41.768 [INFO][4265] cni-plugin/k8s.go 418: Populated endpoint ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0", GenerateName:"calico-apiserver-86dfc84df9-", Namespace:"calico-apiserver", SelfLink:"", UID:"e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86dfc84df9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"calico-apiserver-86dfc84df9-nsjp8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.83.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9b5fd7bc851", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:41.841006 containerd[1544]: 2025-10-13 00:03:41.768 [INFO][4265] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.4/32] ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" Oct 13 00:03:41.841006 containerd[1544]: 2025-10-13 00:03:41.768 [INFO][4265] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9b5fd7bc851 ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" Oct 13 00:03:41.841006 containerd[1544]: 2025-10-13 00:03:41.800 [INFO][4265] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" Oct 13 00:03:41.841113 containerd[1544]: 2025-10-13 00:03:41.804 [INFO][4265] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0", GenerateName:"calico-apiserver-86dfc84df9-", Namespace:"calico-apiserver", SelfLink:"", UID:"e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86dfc84df9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659", Pod:"calico-apiserver-86dfc84df9-nsjp8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.83.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9b5fd7bc851", MAC:"1e:20:b6:5b:de:82", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:41.841165 containerd[1544]: 2025-10-13 00:03:41.835 [INFO][4265] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" Namespace="calico-apiserver" Pod="calico-apiserver-86dfc84df9-nsjp8" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-calico--apiserver--86dfc84df9--nsjp8-eth0" Oct 13 00:03:41.857885 systemd[1]: Started cri-containerd-0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882.scope - libcontainer container 0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882. Oct 13 00:03:41.881702 containerd[1544]: time="2025-10-13T00:03:41.881596203Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:41.886564 containerd[1544]: time="2025-10-13T00:03:41.886466469Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=30823700" Oct 13 00:03:41.888000 containerd[1544]: time="2025-10-13T00:03:41.887848036Z" level=info msg="ImageCreate event name:\"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:41.898988 containerd[1544]: time="2025-10-13T00:03:41.898099121Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:41.902141 containerd[1544]: time="2025-10-13T00:03:41.901047307Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"30823530\" in 3.499633861s" Oct 13 00:03:41.902141 containerd[1544]: time="2025-10-13T00:03:41.901096030Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\"" Oct 13 00:03:41.906238 systemd-networkd[1418]: cali2cb5bc5978d: Link UP Oct 13 00:03:41.908076 containerd[1544]: time="2025-10-13T00:03:41.907299181Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Oct 13 00:03:41.908144 systemd-networkd[1418]: cali2cb5bc5978d: Gained carrier Oct 13 00:03:41.914489 containerd[1544]: time="2025-10-13T00:03:41.912206209Z" level=info msg="CreateContainer within sandbox \"307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Oct 13 00:03:41.918640 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount22913920.mount: Deactivated successfully. Oct 13 00:03:41.945999 containerd[1544]: time="2025-10-13T00:03:41.945721919Z" level=info msg="connecting to shim 74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659" address="unix:///run/containerd/s/7b9e9a8d880053fa3f23f8a45d8c26b58602eb3fa33a13805ab3c5dec0f8d95f" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:41.948286 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1990687782.mount: Deactivated successfully. Oct 13 00:03:41.962221 containerd[1544]: time="2025-10-13T00:03:41.962154794Z" level=info msg="Container 7fe37af1f30a34113fa2c328d87daefadfc7fdb9e2943d456ec737529f0d6576: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:41.981525 containerd[1544]: 2025-10-13 00:03:41.476 [INFO][4285] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0 coredns-668d6bf9bc- kube-system caf89f26-46b1-4e1a-9e57-9e21ff840fc3 801 0 2025-10-13 00:03:00 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 coredns-668d6bf9bc-v4zvg eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2cb5bc5978d [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-" Oct 13 00:03:41.981525 containerd[1544]: 2025-10-13 00:03:41.483 [INFO][4285] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" Oct 13 00:03:41.981525 containerd[1544]: 2025-10-13 00:03:41.637 [INFO][4341] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" HandleID="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Workload="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.638 [INFO][4341] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" HandleID="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Workload="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002d0f20), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459-1-0-3-bea61e6062", "pod":"coredns-668d6bf9bc-v4zvg", "timestamp":"2025-10-13 00:03:41.637733732 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.638 [INFO][4341] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.753 [INFO][4341] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.753 [INFO][4341] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.788 [INFO][4341] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.808 [INFO][4341] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.825 [INFO][4341] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.835 [INFO][4341] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.981921 containerd[1544]: 2025-10-13 00:03:41.844 [INFO][4341] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.983534 containerd[1544]: 2025-10-13 00:03:41.844 [INFO][4341] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.983534 containerd[1544]: 2025-10-13 00:03:41.852 [INFO][4341] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a Oct 13 00:03:41.983534 containerd[1544]: 2025-10-13 00:03:41.864 [INFO][4341] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.983534 containerd[1544]: 2025-10-13 00:03:41.879 [INFO][4341] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.5/26] block=192.168.83.0/26 handle="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.983534 containerd[1544]: 2025-10-13 00:03:41.880 [INFO][4341] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.5/26] handle="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:41.983534 containerd[1544]: 2025-10-13 00:03:41.880 [INFO][4341] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:41.983534 containerd[1544]: 2025-10-13 00:03:41.880 [INFO][4341] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.5/26] IPv6=[] ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" HandleID="k8s-pod-network.6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Workload="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" Oct 13 00:03:41.983702 containerd[1544]: 2025-10-13 00:03:41.885 [INFO][4285] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"caf89f26-46b1-4e1a-9e57-9e21ff840fc3", ResourceVersion:"801", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"coredns-668d6bf9bc-v4zvg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.83.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2cb5bc5978d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:41.983702 containerd[1544]: 2025-10-13 00:03:41.886 [INFO][4285] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.5/32] ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" Oct 13 00:03:41.983702 containerd[1544]: 2025-10-13 00:03:41.886 [INFO][4285] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2cb5bc5978d ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" Oct 13 00:03:41.983702 containerd[1544]: 2025-10-13 00:03:41.909 [INFO][4285] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" Oct 13 00:03:41.983702 containerd[1544]: 2025-10-13 00:03:41.922 [INFO][4285] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"caf89f26-46b1-4e1a-9e57-9e21ff840fc3", ResourceVersion:"801", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a", Pod:"coredns-668d6bf9bc-v4zvg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.83.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2cb5bc5978d", MAC:"1a:a1:16:c2:40:7d", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:41.983702 containerd[1544]: 2025-10-13 00:03:41.959 [INFO][4285] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" Namespace="kube-system" Pod="coredns-668d6bf9bc-v4zvg" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--v4zvg-eth0" Oct 13 00:03:42.038376 containerd[1544]: time="2025-10-13T00:03:42.038284797Z" level=info msg="CreateContainer within sandbox \"307b6336984befc33e8a76b5c3ea75db81388d5675c89d57f374fe88a3768fcc\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"7fe37af1f30a34113fa2c328d87daefadfc7fdb9e2943d456ec737529f0d6576\"" Oct 13 00:03:42.038697 systemd[1]: Started cri-containerd-74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659.scope - libcontainer container 74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659. Oct 13 00:03:42.044976 containerd[1544]: time="2025-10-13T00:03:42.044927250Z" level=info msg="StartContainer for \"7fe37af1f30a34113fa2c328d87daefadfc7fdb9e2943d456ec737529f0d6576\"" Oct 13 00:03:42.102180 containerd[1544]: time="2025-10-13T00:03:42.102128407Z" level=info msg="connecting to shim 7fe37af1f30a34113fa2c328d87daefadfc7fdb9e2943d456ec737529f0d6576" address="unix:///run/containerd/s/fbf5c6ae55aaf257ba4c10d543a75529b712644b7b1e0040084f20b94120d2a4" protocol=ttrpc version=3 Oct 13 00:03:42.102823 containerd[1544]: time="2025-10-13T00:03:42.102747725Z" level=info msg="connecting to shim 6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a" address="unix:///run/containerd/s/c8dc2a909156b3977c05fc575450133f5b29c66467a151088483ac0d450aebb5" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:42.123520 systemd-networkd[1418]: cali613a9a2972e: Link UP Oct 13 00:03:42.126376 systemd-networkd[1418]: cali613a9a2972e: Gained carrier Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.434 [INFO][4256] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0 coredns-668d6bf9bc- kube-system 1ad67c87-3ab9-47d8-91ac-3b7ac8198689 797 0 2025-10-13 00:03:00 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 coredns-668d6bf9bc-hls85 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali613a9a2972e [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.434 [INFO][4256] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.648 [INFO][4321] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" HandleID="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Workload="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.652 [INFO][4321] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" HandleID="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Workload="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000277900), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459-1-0-3-bea61e6062", "pod":"coredns-668d6bf9bc-hls85", "timestamp":"2025-10-13 00:03:41.648940637 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.653 [INFO][4321] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.880 [INFO][4321] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.880 [INFO][4321] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.937 [INFO][4321] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.962 [INFO][4321] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:41.994 [INFO][4321] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.006 [INFO][4321] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.017 [INFO][4321] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.017 [INFO][4321] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.023 [INFO][4321] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6 Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.045 [INFO][4321] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.083 [INFO][4321] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.6/26] block=192.168.83.0/26 handle="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.084 [INFO][4321] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.6/26] handle="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.084 [INFO][4321] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:42.169618 containerd[1544]: 2025-10-13 00:03:42.084 [INFO][4321] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.6/26] IPv6=[] ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" HandleID="k8s-pod-network.8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Workload="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" Oct 13 00:03:42.171205 containerd[1544]: 2025-10-13 00:03:42.106 [INFO][4256] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1ad67c87-3ab9-47d8-91ac-3b7ac8198689", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"coredns-668d6bf9bc-hls85", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.83.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali613a9a2972e", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:42.171205 containerd[1544]: 2025-10-13 00:03:42.106 [INFO][4256] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.6/32] ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" Oct 13 00:03:42.171205 containerd[1544]: 2025-10-13 00:03:42.106 [INFO][4256] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali613a9a2972e ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" Oct 13 00:03:42.171205 containerd[1544]: 2025-10-13 00:03:42.127 [INFO][4256] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" Oct 13 00:03:42.171205 containerd[1544]: 2025-10-13 00:03:42.135 [INFO][4256] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1ad67c87-3ab9-47d8-91ac-3b7ac8198689", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6", Pod:"coredns-668d6bf9bc-hls85", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.83.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali613a9a2972e", MAC:"72:42:ea:8c:58:72", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:42.171205 containerd[1544]: 2025-10-13 00:03:42.164 [INFO][4256] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" Namespace="kube-system" Pod="coredns-668d6bf9bc-hls85" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-coredns--668d6bf9bc--hls85-eth0" Oct 13 00:03:42.170107 systemd[1]: Started cri-containerd-6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a.scope - libcontainer container 6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a. Oct 13 00:03:42.240638 containerd[1544]: time="2025-10-13T00:03:42.239386861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-8cklf,Uid:6a443b6d-f54f-4393-a27f-1337e1a343c9,Namespace:calico-system,Attempt:0,}" Oct 13 00:03:42.259655 systemd-networkd[1418]: cali0ccca2aee67: Link UP Oct 13 00:03:42.262357 systemd-networkd[1418]: cali0ccca2aee67: Gained carrier Oct 13 00:03:42.266934 containerd[1544]: time="2025-10-13T00:03:42.266671637Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-jd8l6,Uid:cd2688c6-7370-41d4-a4a6-1d1ea0156d7e,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882\"" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:41.458 [INFO][4260] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0 csi-node-driver- calico-system c82ba31d-2f24-4bf3-a00d-09a349524d1c 669 0 2025-10-13 00:03:16 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 csi-node-driver-mtmkk eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali0ccca2aee67 [] [] }} ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:41.460 [INFO][4260] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:41.657 [INFO][4337] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" HandleID="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Workload="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:41.657 [INFO][4337] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" HandleID="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Workload="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000102790), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-1-0-3-bea61e6062", "pod":"csi-node-driver-mtmkk", "timestamp":"2025-10-13 00:03:41.657152314 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:41.657 [INFO][4337] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.086 [INFO][4337] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.086 [INFO][4337] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.144 [INFO][4337] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.169 [INFO][4337] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.195 [INFO][4337] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.202 [INFO][4337] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.211 [INFO][4337] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.211 [INFO][4337] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.215 [INFO][4337] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500 Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.221 [INFO][4337] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.234 [INFO][4337] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.7/26] block=192.168.83.0/26 handle="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.234 [INFO][4337] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.7/26] handle="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.234 [INFO][4337] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:42.313018 containerd[1544]: 2025-10-13 00:03:42.234 [INFO][4337] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.7/26] IPv6=[] ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" HandleID="k8s-pod-network.83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Workload="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" Oct 13 00:03:42.314994 containerd[1544]: 2025-10-13 00:03:42.242 [INFO][4260] cni-plugin/k8s.go 418: Populated endpoint ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"c82ba31d-2f24-4bf3-a00d-09a349524d1c", ResourceVersion:"669", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"csi-node-driver-mtmkk", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.83.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali0ccca2aee67", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:42.314994 containerd[1544]: 2025-10-13 00:03:42.244 [INFO][4260] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.7/32] ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" Oct 13 00:03:42.314994 containerd[1544]: 2025-10-13 00:03:42.244 [INFO][4260] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0ccca2aee67 ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" Oct 13 00:03:42.314994 containerd[1544]: 2025-10-13 00:03:42.265 [INFO][4260] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" Oct 13 00:03:42.314994 containerd[1544]: 2025-10-13 00:03:42.281 [INFO][4260] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"c82ba31d-2f24-4bf3-a00d-09a349524d1c", ResourceVersion:"669", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500", Pod:"csi-node-driver-mtmkk", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.83.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali0ccca2aee67", MAC:"e6:b6:f8:3f:98:15", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:42.314994 containerd[1544]: 2025-10-13 00:03:42.304 [INFO][4260] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" Namespace="calico-system" Pod="csi-node-driver-mtmkk" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-csi--node--driver--mtmkk-eth0" Oct 13 00:03:42.383047 systemd[1]: Started cri-containerd-7fe37af1f30a34113fa2c328d87daefadfc7fdb9e2943d456ec737529f0d6576.scope - libcontainer container 7fe37af1f30a34113fa2c328d87daefadfc7fdb9e2943d456ec737529f0d6576. Oct 13 00:03:42.454629 containerd[1544]: time="2025-10-13T00:03:42.454573840Z" level=info msg="connecting to shim 8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6" address="unix:///run/containerd/s/e88399de4c662b26ac8b9ece30a60bdf8fa30c7659a49bc284e1b7678f413436" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:42.497771 containerd[1544]: time="2025-10-13T00:03:42.497364180Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-v4zvg,Uid:caf89f26-46b1-4e1a-9e57-9e21ff840fc3,Namespace:kube-system,Attempt:0,} returns sandbox id \"6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a\"" Oct 13 00:03:42.513595 containerd[1544]: time="2025-10-13T00:03:42.513542586Z" level=info msg="connecting to shim 83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500" address="unix:///run/containerd/s/635cbb919995b2c20411d3711b911662395bf4fd900166950b5c26842bc301ed" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:42.528287 containerd[1544]: time="2025-10-13T00:03:42.528226219Z" level=info msg="CreateContainer within sandbox \"6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 13 00:03:42.601130 systemd[1]: Started cri-containerd-83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500.scope - libcontainer container 83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500. Oct 13 00:03:42.607108 containerd[1544]: time="2025-10-13T00:03:42.607036119Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86dfc84df9-nsjp8,Uid:e9e6f72d-27e8-4bcc-ae87-2bfe7c24625f,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659\"" Oct 13 00:03:42.631624 systemd[1]: Started cri-containerd-8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6.scope - libcontainer container 8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6. Oct 13 00:03:42.651209 containerd[1544]: time="2025-10-13T00:03:42.651134141Z" level=info msg="Container b5b12b7ad5be81467890f93a390a7b163d25d95a3bc5e6ad94a0df8cc309bfab: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:42.665675 containerd[1544]: time="2025-10-13T00:03:42.665581999Z" level=info msg="CreateContainer within sandbox \"6288035f98e89121ce05ce598bcaeefbe1e5ae318b01b6e33d076921d2e7a49a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b5b12b7ad5be81467890f93a390a7b163d25d95a3bc5e6ad94a0df8cc309bfab\"" Oct 13 00:03:42.667137 containerd[1544]: time="2025-10-13T00:03:42.666925883Z" level=info msg="StartContainer for \"b5b12b7ad5be81467890f93a390a7b163d25d95a3bc5e6ad94a0df8cc309bfab\"" Oct 13 00:03:42.670248 containerd[1544]: time="2025-10-13T00:03:42.670157484Z" level=info msg="connecting to shim b5b12b7ad5be81467890f93a390a7b163d25d95a3bc5e6ad94a0df8cc309bfab" address="unix:///run/containerd/s/c8dc2a909156b3977c05fc575450133f5b29c66467a151088483ac0d450aebb5" protocol=ttrpc version=3 Oct 13 00:03:42.715658 systemd[1]: Started cri-containerd-b5b12b7ad5be81467890f93a390a7b163d25d95a3bc5e6ad94a0df8cc309bfab.scope - libcontainer container b5b12b7ad5be81467890f93a390a7b163d25d95a3bc5e6ad94a0df8cc309bfab. Oct 13 00:03:42.751865 containerd[1544]: time="2025-10-13T00:03:42.750637167Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hls85,Uid:1ad67c87-3ab9-47d8-91ac-3b7ac8198689,Namespace:kube-system,Attempt:0,} returns sandbox id \"8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6\"" Oct 13 00:03:42.774021 containerd[1544]: time="2025-10-13T00:03:42.773960617Z" level=info msg="CreateContainer within sandbox \"8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 13 00:03:42.819674 containerd[1544]: time="2025-10-13T00:03:42.819635897Z" level=info msg="Container effed4a66a580175ba9ea6e17f9bf7c0b5adc9ed94df90cc0ce1a353f71f40b8: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:42.834151 containerd[1544]: time="2025-10-13T00:03:42.834050793Z" level=info msg="StartContainer for \"7fe37af1f30a34113fa2c328d87daefadfc7fdb9e2943d456ec737529f0d6576\" returns successfully" Oct 13 00:03:42.850887 systemd-networkd[1418]: cali7e5d3ccec56: Link UP Oct 13 00:03:42.853957 systemd-networkd[1418]: cali7e5d3ccec56: Gained carrier Oct 13 00:03:42.854566 containerd[1544]: time="2025-10-13T00:03:42.854529187Z" level=info msg="StartContainer for \"b5b12b7ad5be81467890f93a390a7b163d25d95a3bc5e6ad94a0df8cc309bfab\" returns successfully" Oct 13 00:03:42.860331 containerd[1544]: time="2025-10-13T00:03:42.860196139Z" level=info msg="CreateContainer within sandbox \"8a6bb90a7b36c8fcce15c75884b597bc31eb3bc8342a4066ebe9afe0a50be9a6\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"effed4a66a580175ba9ea6e17f9bf7c0b5adc9ed94df90cc0ce1a353f71f40b8\"" Oct 13 00:03:42.864798 containerd[1544]: time="2025-10-13T00:03:42.864287713Z" level=info msg="StartContainer for \"effed4a66a580175ba9ea6e17f9bf7c0b5adc9ed94df90cc0ce1a353f71f40b8\"" Oct 13 00:03:42.867463 containerd[1544]: time="2025-10-13T00:03:42.866254196Z" level=info msg="connecting to shim effed4a66a580175ba9ea6e17f9bf7c0b5adc9ed94df90cc0ce1a353f71f40b8" address="unix:///run/containerd/s/e88399de4c662b26ac8b9ece30a60bdf8fa30c7659a49bc284e1b7678f413436" protocol=ttrpc version=3 Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.469 [INFO][4539] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0 goldmane-54d579b49d- calico-system 6a443b6d-f54f-4393-a27f-1337e1a343c9 803 0 2025-10-13 00:03:17 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4459-1-0-3-bea61e6062 goldmane-54d579b49d-8cklf eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali7e5d3ccec56 [] [] }} ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.469 [INFO][4539] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.643 [INFO][4604] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" HandleID="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Workload="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.643 [INFO][4604] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" HandleID="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Workload="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003ac3d0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459-1-0-3-bea61e6062", "pod":"goldmane-54d579b49d-8cklf", "timestamp":"2025-10-13 00:03:42.643205688 +0000 UTC"}, Hostname:"ci-4459-1-0-3-bea61e6062", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.643 [INFO][4604] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.643 [INFO][4604] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.644 [INFO][4604] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459-1-0-3-bea61e6062' Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.665 [INFO][4604] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.685 [INFO][4604] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.708 [INFO][4604] ipam/ipam.go 511: Trying affinity for 192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.716 [INFO][4604] ipam/ipam.go 158: Attempting to load block cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.730 [INFO][4604] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.83.0/26 host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.731 [INFO][4604] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.83.0/26 handle="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.750 [INFO][4604] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6 Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.770 [INFO][4604] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.83.0/26 handle="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.802 [INFO][4604] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.83.8/26] block=192.168.83.0/26 handle="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.802 [INFO][4604] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.83.8/26] handle="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" host="ci-4459-1-0-3-bea61e6062" Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.804 [INFO][4604] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 00:03:42.890467 containerd[1544]: 2025-10-13 00:03:42.805 [INFO][4604] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.83.8/26] IPv6=[] ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" HandleID="k8s-pod-network.b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Workload="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" Oct 13 00:03:42.892273 containerd[1544]: 2025-10-13 00:03:42.830 [INFO][4539] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"6a443b6d-f54f-4393-a27f-1337e1a343c9", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"", Pod:"goldmane-54d579b49d-8cklf", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.83.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7e5d3ccec56", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:42.892273 containerd[1544]: 2025-10-13 00:03:42.831 [INFO][4539] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.83.8/32] ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" Oct 13 00:03:42.892273 containerd[1544]: 2025-10-13 00:03:42.831 [INFO][4539] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7e5d3ccec56 ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" Oct 13 00:03:42.892273 containerd[1544]: 2025-10-13 00:03:42.853 [INFO][4539] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" Oct 13 00:03:42.892273 containerd[1544]: 2025-10-13 00:03:42.860 [INFO][4539] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"6a443b6d-f54f-4393-a27f-1337e1a343c9", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 0, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459-1-0-3-bea61e6062", ContainerID:"b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6", Pod:"goldmane-54d579b49d-8cklf", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.83.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7e5d3ccec56", MAC:"7a:96:a8:ee:a8:fd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 00:03:42.892273 containerd[1544]: 2025-10-13 00:03:42.885 [INFO][4539] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" Namespace="calico-system" Pod="goldmane-54d579b49d-8cklf" WorkloadEndpoint="ci--4459--1--0--3--bea61e6062-k8s-goldmane--54d579b49d--8cklf-eth0" Oct 13 00:03:42.911221 systemd-networkd[1418]: cali20d2d13c1b6: Gained IPv6LL Oct 13 00:03:42.983634 systemd[1]: Started cri-containerd-effed4a66a580175ba9ea6e17f9bf7c0b5adc9ed94df90cc0ce1a353f71f40b8.scope - libcontainer container effed4a66a580175ba9ea6e17f9bf7c0b5adc9ed94df90cc0ce1a353f71f40b8. Oct 13 00:03:42.998652 containerd[1544]: time="2025-10-13T00:03:42.998312286Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mtmkk,Uid:c82ba31d-2f24-4bf3-a00d-09a349524d1c,Namespace:calico-system,Attempt:0,} returns sandbox id \"83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500\"" Oct 13 00:03:43.010575 containerd[1544]: time="2025-10-13T00:03:43.010395391Z" level=info msg="connecting to shim b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6" address="unix:///run/containerd/s/d1781f4d087af3958ceb199b46ea6515254eb389a37d1f97aa3946d509b20c68" namespace=k8s.io protocol=ttrpc version=3 Oct 13 00:03:43.067594 systemd[1]: Started cri-containerd-b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6.scope - libcontainer container b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6. Oct 13 00:03:43.113951 containerd[1544]: time="2025-10-13T00:03:43.113913152Z" level=info msg="StartContainer for \"effed4a66a580175ba9ea6e17f9bf7c0b5adc9ed94df90cc0ce1a353f71f40b8\" returns successfully" Oct 13 00:03:43.147164 containerd[1544]: time="2025-10-13T00:03:43.147113592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-8cklf,Uid:6a443b6d-f54f-4393-a27f-1337e1a343c9,Namespace:calico-system,Attempt:0,} returns sandbox id \"b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6\"" Oct 13 00:03:43.166607 systemd-networkd[1418]: cali613a9a2972e: Gained IPv6LL Oct 13 00:03:43.486722 systemd-networkd[1418]: cali9b5fd7bc851: Gained IPv6LL Oct 13 00:03:43.641741 kubelet[2740]: I1013 00:03:43.641682 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-57df5ddf6b-bw8z2" podStartSLOduration=3.25370749 podStartE2EDuration="8.641666101s" podCreationTimestamp="2025-10-13 00:03:35 +0000 UTC" firstStartedPulling="2025-10-13 00:03:36.51727468 +0000 UTC m=+43.444137779" lastFinishedPulling="2025-10-13 00:03:41.905233291 +0000 UTC m=+48.832096390" observedRunningTime="2025-10-13 00:03:43.639390241 +0000 UTC m=+50.566253340" watchObservedRunningTime="2025-10-13 00:03:43.641666101 +0000 UTC m=+50.568529200" Oct 13 00:03:43.698724 kubelet[2740]: I1013 00:03:43.698559 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-hls85" podStartSLOduration=43.698536676 podStartE2EDuration="43.698536676s" podCreationTimestamp="2025-10-13 00:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 00:03:43.655127368 +0000 UTC m=+50.581990587" watchObservedRunningTime="2025-10-13 00:03:43.698536676 +0000 UTC m=+50.625399775" Oct 13 00:03:43.742597 systemd-networkd[1418]: cali2cb5bc5978d: Gained IPv6LL Oct 13 00:03:43.999165 systemd-networkd[1418]: cali0ccca2aee67: Gained IPv6LL Oct 13 00:03:44.703959 systemd-networkd[1418]: cali7e5d3ccec56: Gained IPv6LL Oct 13 00:03:45.113399 containerd[1544]: time="2025-10-13T00:03:45.113316501Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:45.114756 containerd[1544]: time="2025-10-13T00:03:45.114682023Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=48134957" Oct 13 00:03:45.115822 containerd[1544]: time="2025-10-13T00:03:45.115763328Z" level=info msg="ImageCreate event name:\"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:45.119360 containerd[1544]: time="2025-10-13T00:03:45.119296221Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:45.120341 containerd[1544]: time="2025-10-13T00:03:45.119770889Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"49504166\" in 3.212430346s" Oct 13 00:03:45.120341 containerd[1544]: time="2025-10-13T00:03:45.119824252Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\"" Oct 13 00:03:45.122084 containerd[1544]: time="2025-10-13T00:03:45.122016744Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Oct 13 00:03:45.146024 containerd[1544]: time="2025-10-13T00:03:45.145928702Z" level=info msg="CreateContainer within sandbox \"fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Oct 13 00:03:45.158472 containerd[1544]: time="2025-10-13T00:03:45.158426813Z" level=info msg="Container 143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:45.169610 containerd[1544]: time="2025-10-13T00:03:45.169376272Z" level=info msg="CreateContainer within sandbox \"fdcb89f4196bbbef86adfcfbf379ed5af80917140644bf1cc5d7968eb7e1f933\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\"" Oct 13 00:03:45.171311 containerd[1544]: time="2025-10-13T00:03:45.170541382Z" level=info msg="StartContainer for \"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\"" Oct 13 00:03:45.173780 containerd[1544]: time="2025-10-13T00:03:45.173733614Z" level=info msg="connecting to shim 143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184" address="unix:///run/containerd/s/b8f314fd7661da5a49ebb1cef9d8c6b06146a540f254b3de0bac789f4347ed63" protocol=ttrpc version=3 Oct 13 00:03:45.200642 systemd[1]: Started cri-containerd-143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184.scope - libcontainer container 143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184. Oct 13 00:03:45.253970 containerd[1544]: time="2025-10-13T00:03:45.253877433Z" level=info msg="StartContainer for \"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" returns successfully" Oct 13 00:03:45.664290 kubelet[2740]: I1013 00:03:45.663657 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-76b7656b99-t5tjj" podStartSLOduration=24.230414291 podStartE2EDuration="29.663587148s" podCreationTimestamp="2025-10-13 00:03:16 +0000 UTC" firstStartedPulling="2025-10-13 00:03:39.688027158 +0000 UTC m=+46.614890257" lastFinishedPulling="2025-10-13 00:03:45.121200015 +0000 UTC m=+52.048063114" observedRunningTime="2025-10-13 00:03:45.663195244 +0000 UTC m=+52.590058343" watchObservedRunningTime="2025-10-13 00:03:45.663587148 +0000 UTC m=+52.590471928" Oct 13 00:03:45.665039 kubelet[2740]: I1013 00:03:45.664930 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-v4zvg" podStartSLOduration=45.664883146 podStartE2EDuration="45.664883146s" podCreationTimestamp="2025-10-13 00:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 00:03:43.699839396 +0000 UTC m=+50.626702495" watchObservedRunningTime="2025-10-13 00:03:45.664883146 +0000 UTC m=+52.591746245" Oct 13 00:03:45.691311 containerd[1544]: time="2025-10-13T00:03:45.691209489Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"e7f14cb39938abaf7e5aa67e2377fca4697417b1e9c71b646140b6a0331c54d9\" pid:4886 exited_at:{seconds:1760313825 nanos:690774463}" Oct 13 00:03:48.344213 containerd[1544]: time="2025-10-13T00:03:48.343059685Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:48.344213 containerd[1544]: time="2025-10-13T00:03:48.344158670Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=44530807" Oct 13 00:03:48.345143 containerd[1544]: time="2025-10-13T00:03:48.345035561Z" level=info msg="ImageCreate event name:\"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:48.348224 containerd[1544]: time="2025-10-13T00:03:48.348167944Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:48.349073 containerd[1544]: time="2025-10-13T00:03:48.349042395Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"45900064\" in 3.226963927s" Oct 13 00:03:48.349233 containerd[1544]: time="2025-10-13T00:03:48.349212685Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\"" Oct 13 00:03:48.350580 containerd[1544]: time="2025-10-13T00:03:48.350532922Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Oct 13 00:03:48.351848 containerd[1544]: time="2025-10-13T00:03:48.351806877Z" level=info msg="CreateContainer within sandbox \"0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Oct 13 00:03:48.364781 containerd[1544]: time="2025-10-13T00:03:48.364728112Z" level=info msg="Container b0ebc711f3f38dc5c281082dd20c5df8e30f925512f56e74faa354457a112cc4: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:48.370579 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount467144481.mount: Deactivated successfully. Oct 13 00:03:48.381645 containerd[1544]: time="2025-10-13T00:03:48.381591977Z" level=info msg="CreateContainer within sandbox \"0d681ba3c49a1daf3fd52575c4b6ee74bb3e9b54ef100457374eae1f0ca4e882\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"b0ebc711f3f38dc5c281082dd20c5df8e30f925512f56e74faa354457a112cc4\"" Oct 13 00:03:48.382958 containerd[1544]: time="2025-10-13T00:03:48.382927295Z" level=info msg="StartContainer for \"b0ebc711f3f38dc5c281082dd20c5df8e30f925512f56e74faa354457a112cc4\"" Oct 13 00:03:48.385804 containerd[1544]: time="2025-10-13T00:03:48.385720739Z" level=info msg="connecting to shim b0ebc711f3f38dc5c281082dd20c5df8e30f925512f56e74faa354457a112cc4" address="unix:///run/containerd/s/e17e363542aede78e81f6544e8f4ce021bc5a7065a82552e161522fbedbbef75" protocol=ttrpc version=3 Oct 13 00:03:48.425175 systemd[1]: Started cri-containerd-b0ebc711f3f38dc5c281082dd20c5df8e30f925512f56e74faa354457a112cc4.scope - libcontainer container b0ebc711f3f38dc5c281082dd20c5df8e30f925512f56e74faa354457a112cc4. Oct 13 00:03:48.474628 containerd[1544]: time="2025-10-13T00:03:48.474386600Z" level=info msg="StartContainer for \"b0ebc711f3f38dc5c281082dd20c5df8e30f925512f56e74faa354457a112cc4\" returns successfully" Oct 13 00:03:48.759502 containerd[1544]: time="2025-10-13T00:03:48.758463803Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:48.760013 containerd[1544]: time="2025-10-13T00:03:48.759982692Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=77" Oct 13 00:03:48.764230 containerd[1544]: time="2025-10-13T00:03:48.764048209Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"45900064\" in 413.470805ms" Oct 13 00:03:48.764230 containerd[1544]: time="2025-10-13T00:03:48.764098932Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\"" Oct 13 00:03:48.767695 containerd[1544]: time="2025-10-13T00:03:48.767190393Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Oct 13 00:03:48.771467 containerd[1544]: time="2025-10-13T00:03:48.769912272Z" level=info msg="CreateContainer within sandbox \"74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Oct 13 00:03:48.814629 containerd[1544]: time="2025-10-13T00:03:48.814577802Z" level=info msg="Container 811fe00c22416e3ffb477e7af2ceb0a901c4df822f084adac9c237e1415918e2: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:48.840205 containerd[1544]: time="2025-10-13T00:03:48.840103734Z" level=info msg="CreateContainer within sandbox \"74f488e2740e7b9e76ddba643cc082876a1d692a693e64034714314561bc8659\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"811fe00c22416e3ffb477e7af2ceb0a901c4df822f084adac9c237e1415918e2\"" Oct 13 00:03:48.841612 containerd[1544]: time="2025-10-13T00:03:48.841298884Z" level=info msg="StartContainer for \"811fe00c22416e3ffb477e7af2ceb0a901c4df822f084adac9c237e1415918e2\"" Oct 13 00:03:48.843550 containerd[1544]: time="2025-10-13T00:03:48.843456010Z" level=info msg="connecting to shim 811fe00c22416e3ffb477e7af2ceb0a901c4df822f084adac9c237e1415918e2" address="unix:///run/containerd/s/7b9e9a8d880053fa3f23f8a45d8c26b58602eb3fa33a13805ab3c5dec0f8d95f" protocol=ttrpc version=3 Oct 13 00:03:48.875657 systemd[1]: Started cri-containerd-811fe00c22416e3ffb477e7af2ceb0a901c4df822f084adac9c237e1415918e2.scope - libcontainer container 811fe00c22416e3ffb477e7af2ceb0a901c4df822f084adac9c237e1415918e2. Oct 13 00:03:48.932718 containerd[1544]: time="2025-10-13T00:03:48.932619581Z" level=info msg="StartContainer for \"811fe00c22416e3ffb477e7af2ceb0a901c4df822f084adac9c237e1415918e2\" returns successfully" Oct 13 00:03:49.668607 kubelet[2740]: I1013 00:03:49.667476 2740 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 00:03:49.702427 kubelet[2740]: I1013 00:03:49.702259 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-86dfc84df9-jd8l6" podStartSLOduration=32.635195014 podStartE2EDuration="38.702222173s" podCreationTimestamp="2025-10-13 00:03:11 +0000 UTC" firstStartedPulling="2025-10-13 00:03:42.283167463 +0000 UTC m=+49.210030562" lastFinishedPulling="2025-10-13 00:03:48.350194622 +0000 UTC m=+55.277057721" observedRunningTime="2025-10-13 00:03:48.689553776 +0000 UTC m=+55.616416875" watchObservedRunningTime="2025-10-13 00:03:49.702222173 +0000 UTC m=+56.629085232" Oct 13 00:03:49.703141 kubelet[2740]: I1013 00:03:49.703090 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-86dfc84df9-nsjp8" podStartSLOduration=32.562988715 podStartE2EDuration="38.703076502s" podCreationTimestamp="2025-10-13 00:03:11 +0000 UTC" firstStartedPulling="2025-10-13 00:03:42.626882313 +0000 UTC m=+49.553745412" lastFinishedPulling="2025-10-13 00:03:48.76697014 +0000 UTC m=+55.693833199" observedRunningTime="2025-10-13 00:03:49.693358259 +0000 UTC m=+56.620221358" watchObservedRunningTime="2025-10-13 00:03:49.703076502 +0000 UTC m=+56.629939601" Oct 13 00:03:50.423884 containerd[1544]: time="2025-10-13T00:03:50.423838995Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:50.425755 containerd[1544]: time="2025-10-13T00:03:50.425711863Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8227489" Oct 13 00:03:50.427485 containerd[1544]: time="2025-10-13T00:03:50.427389319Z" level=info msg="ImageCreate event name:\"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:50.433463 containerd[1544]: time="2025-10-13T00:03:50.432876714Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:50.433721 containerd[1544]: time="2025-10-13T00:03:50.433681681Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"9596730\" in 1.666448005s" Oct 13 00:03:50.433770 containerd[1544]: time="2025-10-13T00:03:50.433719203Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\"" Oct 13 00:03:50.435858 containerd[1544]: time="2025-10-13T00:03:50.435608431Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Oct 13 00:03:50.437283 containerd[1544]: time="2025-10-13T00:03:50.437240445Z" level=info msg="CreateContainer within sandbox \"83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Oct 13 00:03:50.467438 containerd[1544]: time="2025-10-13T00:03:50.467132884Z" level=info msg="Container 146b76cda7540b75eb412fc087fbeba19089788c4f1fe1394b66ef141a672b9d: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:50.482350 containerd[1544]: time="2025-10-13T00:03:50.482273194Z" level=info msg="CreateContainer within sandbox \"83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"146b76cda7540b75eb412fc087fbeba19089788c4f1fe1394b66ef141a672b9d\"" Oct 13 00:03:50.485702 containerd[1544]: time="2025-10-13T00:03:50.484636730Z" level=info msg="StartContainer for \"146b76cda7540b75eb412fc087fbeba19089788c4f1fe1394b66ef141a672b9d\"" Oct 13 00:03:50.487441 containerd[1544]: time="2025-10-13T00:03:50.486925862Z" level=info msg="connecting to shim 146b76cda7540b75eb412fc087fbeba19089788c4f1fe1394b66ef141a672b9d" address="unix:///run/containerd/s/635cbb919995b2c20411d3711b911662395bf4fd900166950b5c26842bc301ed" protocol=ttrpc version=3 Oct 13 00:03:50.516629 systemd[1]: Started cri-containerd-146b76cda7540b75eb412fc087fbeba19089788c4f1fe1394b66ef141a672b9d.scope - libcontainer container 146b76cda7540b75eb412fc087fbeba19089788c4f1fe1394b66ef141a672b9d. Oct 13 00:03:50.583096 containerd[1544]: time="2025-10-13T00:03:50.583024986Z" level=info msg="StartContainer for \"146b76cda7540b75eb412fc087fbeba19089788c4f1fe1394b66ef141a672b9d\" returns successfully" Oct 13 00:03:53.207470 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2165586652.mount: Deactivated successfully. Oct 13 00:03:53.693669 containerd[1544]: time="2025-10-13T00:03:53.693595845Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:53.695586 containerd[1544]: time="2025-10-13T00:03:53.695226473Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=61845332" Oct 13 00:03:53.696865 containerd[1544]: time="2025-10-13T00:03:53.696791945Z" level=info msg="ImageCreate event name:\"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:53.701481 containerd[1544]: time="2025-10-13T00:03:53.701259302Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:53.702854 containerd[1544]: time="2025-10-13T00:03:53.702257861Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"61845178\" in 3.266605029s" Oct 13 00:03:53.702854 containerd[1544]: time="2025-10-13T00:03:53.702302458Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\"" Oct 13 00:03:53.705658 containerd[1544]: time="2025-10-13T00:03:53.705624988Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Oct 13 00:03:53.707229 containerd[1544]: time="2025-10-13T00:03:53.707186221Z" level=info msg="CreateContainer within sandbox \"b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Oct 13 00:03:53.721699 containerd[1544]: time="2025-10-13T00:03:53.721652285Z" level=info msg="Container e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:53.734334 containerd[1544]: time="2025-10-13T00:03:53.734252461Z" level=info msg="CreateContainer within sandbox \"b4b0920b7c2d21c27d38236c184133f8b34808574e95128a2aa52afe44a5f6d6\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\"" Oct 13 00:03:53.735832 containerd[1544]: time="2025-10-13T00:03:53.735777937Z" level=info msg="StartContainer for \"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\"" Oct 13 00:03:53.740511 containerd[1544]: time="2025-10-13T00:03:53.739616825Z" level=info msg="connecting to shim e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace" address="unix:///run/containerd/s/d1781f4d087af3958ceb199b46ea6515254eb389a37d1f97aa3946d509b20c68" protocol=ttrpc version=3 Oct 13 00:03:53.769683 systemd[1]: Started cri-containerd-e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace.scope - libcontainer container e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace. Oct 13 00:03:53.833662 containerd[1544]: time="2025-10-13T00:03:53.833541871Z" level=info msg="StartContainer for \"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" returns successfully" Oct 13 00:03:54.739542 kubelet[2740]: I1013 00:03:54.738717 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-8cklf" podStartSLOduration=27.187295987 podStartE2EDuration="37.738692069s" podCreationTimestamp="2025-10-13 00:03:17 +0000 UTC" firstStartedPulling="2025-10-13 00:03:43.152245867 +0000 UTC m=+50.079108966" lastFinishedPulling="2025-10-13 00:03:53.703641949 +0000 UTC m=+60.630505048" observedRunningTime="2025-10-13 00:03:54.731744928 +0000 UTC m=+61.658608027" watchObservedRunningTime="2025-10-13 00:03:54.738692069 +0000 UTC m=+61.665555168" Oct 13 00:03:54.844861 containerd[1544]: time="2025-10-13T00:03:54.844810430Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"65e3a753bcac2e6cb7119ed586fdcb5acf185a03c9445f9d5c1efa735ad3f15b\" pid:5090 exit_status:1 exited_at:{seconds:1760313834 nanos:844055209}" Oct 13 00:03:55.665440 containerd[1544]: time="2025-10-13T00:03:55.665340308Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:55.668245 containerd[1544]: time="2025-10-13T00:03:55.668153979Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=13761208" Oct 13 00:03:55.674457 containerd[1544]: time="2025-10-13T00:03:55.674371998Z" level=info msg="ImageCreate event name:\"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:55.676909 containerd[1544]: time="2025-10-13T00:03:55.676857694Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 00:03:55.677495 containerd[1544]: time="2025-10-13T00:03:55.677459049Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"15130401\" in 1.9716128s" Oct 13 00:03:55.677554 containerd[1544]: time="2025-10-13T00:03:55.677495327Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\"" Oct 13 00:03:55.681765 containerd[1544]: time="2025-10-13T00:03:55.681724973Z" level=info msg="CreateContainer within sandbox \"83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Oct 13 00:03:55.693443 containerd[1544]: time="2025-10-13T00:03:55.691546605Z" level=info msg="Container 4b47816dde12838e8a1bae4a565757e99fa5c4f4aa7b99e10e1b27a758591f34: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:03:55.702192 containerd[1544]: time="2025-10-13T00:03:55.702149939Z" level=info msg="CreateContainer within sandbox \"83946f80a787562e45a21a43558219f6dbec86f8739061d4d046e453ab16c500\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"4b47816dde12838e8a1bae4a565757e99fa5c4f4aa7b99e10e1b27a758591f34\"" Oct 13 00:03:55.703207 containerd[1544]: time="2025-10-13T00:03:55.703185783Z" level=info msg="StartContainer for \"4b47816dde12838e8a1bae4a565757e99fa5c4f4aa7b99e10e1b27a758591f34\"" Oct 13 00:03:55.705020 containerd[1544]: time="2025-10-13T00:03:55.704988089Z" level=info msg="connecting to shim 4b47816dde12838e8a1bae4a565757e99fa5c4f4aa7b99e10e1b27a758591f34" address="unix:///run/containerd/s/635cbb919995b2c20411d3711b911662395bf4fd900166950b5c26842bc301ed" protocol=ttrpc version=3 Oct 13 00:03:55.739146 systemd[1]: Started cri-containerd-4b47816dde12838e8a1bae4a565757e99fa5c4f4aa7b99e10e1b27a758591f34.scope - libcontainer container 4b47816dde12838e8a1bae4a565757e99fa5c4f4aa7b99e10e1b27a758591f34. Oct 13 00:03:55.805671 containerd[1544]: time="2025-10-13T00:03:55.805631230Z" level=info msg="StartContainer for \"4b47816dde12838e8a1bae4a565757e99fa5c4f4aa7b99e10e1b27a758591f34\" returns successfully" Oct 13 00:03:55.854698 containerd[1544]: time="2025-10-13T00:03:55.854632878Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"b3c9b542520438ec84aeab0b5b6d2f0126332c354eb645246797079d8bfc6ed2\" pid:5138 exit_status:1 exited_at:{seconds:1760313835 nanos:854196590}" Oct 13 00:03:56.362655 kubelet[2740]: I1013 00:03:56.362546 2740 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Oct 13 00:03:56.365941 kubelet[2740]: I1013 00:03:56.365904 2740 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Oct 13 00:03:56.827195 containerd[1544]: time="2025-10-13T00:03:56.827026374Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"6d94edc41887ff4e91515b11d8396852c30a57a0b6cabb5f38519fba184a0b4c\" pid:5175 exit_status:1 exited_at:{seconds:1760313836 nanos:826509531}" Oct 13 00:04:07.596085 containerd[1544]: time="2025-10-13T00:04:07.595990911Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"828951e405745a0b9e609886b53e76ee1f5b6c6ef8699de3633ba890718a77e4\" pid:5210 exited_at:{seconds:1760313847 nanos:595691442}" Oct 13 00:04:07.624297 kubelet[2740]: I1013 00:04:07.624199 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-mtmkk" podStartSLOduration=38.948570416 podStartE2EDuration="51.624174717s" podCreationTimestamp="2025-10-13 00:03:16 +0000 UTC" firstStartedPulling="2025-10-13 00:03:43.002755082 +0000 UTC m=+49.929618141" lastFinishedPulling="2025-10-13 00:03:55.678359343 +0000 UTC m=+62.605222442" observedRunningTime="2025-10-13 00:03:56.763094254 +0000 UTC m=+63.689957353" watchObservedRunningTime="2025-10-13 00:04:07.624174717 +0000 UTC m=+74.551037816" Oct 13 00:04:15.820852 containerd[1544]: time="2025-10-13T00:04:15.820796230Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"f24e32b52b74bec2003d3772e702067912a14a088d0ba8f54e130da28207d737\" pid:5234 exited_at:{seconds:1760313855 nanos:820272722}" Oct 13 00:04:21.033225 containerd[1544]: time="2025-10-13T00:04:21.033160002Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"606411b5a467b51f911f714126f218bc78a761db4eef7981ff29857a828aa354\" pid:5263 exited_at:{seconds:1760313861 nanos:31887218}" Oct 13 00:04:22.086333 kubelet[2740]: I1013 00:04:22.086282 2740 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 00:04:26.852364 containerd[1544]: time="2025-10-13T00:04:26.852302175Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"569ec2b270adb144075441989219ed7aa77684c97ac28c332c0634b4cc269384\" pid:5291 exited_at:{seconds:1760313866 nanos:851940257}" Oct 13 00:04:37.600200 containerd[1544]: time="2025-10-13T00:04:37.600139835Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"6ef80476ad782d8457cca6d01821989ed378d062e5b2bbe9211a13c9eb5065a6\" pid:5318 exited_at:{seconds:1760313877 nanos:599805673}" Oct 13 00:04:38.208657 update_engine[1522]: I20251013 00:04:38.208581 1522 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Oct 13 00:04:38.208657 update_engine[1522]: I20251013 00:04:38.208645 1522 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Oct 13 00:04:38.209044 update_engine[1522]: I20251013 00:04:38.208931 1522 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Oct 13 00:04:38.212031 update_engine[1522]: I20251013 00:04:38.211673 1522 omaha_request_params.cc:62] Current group set to beta Oct 13 00:04:38.212898 update_engine[1522]: I20251013 00:04:38.212675 1522 update_attempter.cc:499] Already updated boot flags. Skipping. Oct 13 00:04:38.212898 update_engine[1522]: I20251013 00:04:38.212720 1522 update_attempter.cc:643] Scheduling an action processor start. Oct 13 00:04:38.212898 update_engine[1522]: I20251013 00:04:38.212744 1522 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Oct 13 00:04:38.220540 update_engine[1522]: I20251013 00:04:38.220486 1522 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Oct 13 00:04:38.220669 update_engine[1522]: I20251013 00:04:38.220593 1522 omaha_request_action.cc:271] Posting an Omaha request to disabled Oct 13 00:04:38.220669 update_engine[1522]: I20251013 00:04:38.220602 1522 omaha_request_action.cc:272] Request: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: Oct 13 00:04:38.220669 update_engine[1522]: I20251013 00:04:38.220607 1522 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Oct 13 00:04:38.226452 update_engine[1522]: I20251013 00:04:38.224707 1522 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Oct 13 00:04:38.226452 update_engine[1522]: I20251013 00:04:38.226302 1522 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Oct 13 00:04:38.227187 locksmithd[1564]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Oct 13 00:04:38.227481 update_engine[1522]: E20251013 00:04:38.227207 1522 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Oct 13 00:04:38.227481 update_engine[1522]: I20251013 00:04:38.227278 1522 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Oct 13 00:04:45.694666 containerd[1544]: time="2025-10-13T00:04:45.694527308Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"886e500e480a0d0784c46ae3e1eb409608f3ead6b66fce8954ffce4084809a31\" pid:5343 exited_at:{seconds:1760313885 nanos:693624017}" Oct 13 00:04:48.175719 update_engine[1522]: I20251013 00:04:48.175626 1522 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Oct 13 00:04:48.176168 update_engine[1522]: I20251013 00:04:48.175740 1522 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Oct 13 00:04:48.176168 update_engine[1522]: I20251013 00:04:48.176105 1522 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Oct 13 00:04:48.176588 update_engine[1522]: E20251013 00:04:48.176548 1522 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Oct 13 00:04:48.176662 update_engine[1522]: I20251013 00:04:48.176636 1522 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Oct 13 00:04:51.418703 containerd[1544]: time="2025-10-13T00:04:51.418655856Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"d69a28246adaa91f166313e1533ca8029db123f7a4f5744d3c5fe01bd3e8332d\" pid:5368 exited_at:{seconds:1760313891 nanos:417787002}" Oct 13 00:04:56.810215 containerd[1544]: time="2025-10-13T00:04:56.810165603Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"cb717d0c0e9547201294d995287d07a623c326b0f36df1321a2b035edc70bc4a\" pid:5392 exited_at:{seconds:1760313896 nanos:809781555}" Oct 13 00:04:58.176835 update_engine[1522]: I20251013 00:04:58.176085 1522 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Oct 13 00:04:58.176835 update_engine[1522]: I20251013 00:04:58.176196 1522 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Oct 13 00:04:58.176835 update_engine[1522]: I20251013 00:04:58.176613 1522 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Oct 13 00:04:58.177248 update_engine[1522]: E20251013 00:04:58.177086 1522 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Oct 13 00:04:58.177248 update_engine[1522]: I20251013 00:04:58.177156 1522 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Oct 13 00:05:07.593431 containerd[1544]: time="2025-10-13T00:05:07.593299309Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"3cc9918ea3cd70335eb939a35c6c6e40c6ae5c6dec1abf7f572346585b3e7a93\" pid:5423 exited_at:{seconds:1760313907 nanos:592952101}" Oct 13 00:05:08.176308 update_engine[1522]: I20251013 00:05:08.176201 1522 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Oct 13 00:05:08.176763 update_engine[1522]: I20251013 00:05:08.176322 1522 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Oct 13 00:05:08.176880 update_engine[1522]: I20251013 00:05:08.176811 1522 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Oct 13 00:05:08.177491 update_engine[1522]: E20251013 00:05:08.177274 1522 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Oct 13 00:05:08.177491 update_engine[1522]: I20251013 00:05:08.177381 1522 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Oct 13 00:05:08.177491 update_engine[1522]: I20251013 00:05:08.177395 1522 omaha_request_action.cc:617] Omaha request response: Oct 13 00:05:08.177633 update_engine[1522]: E20251013 00:05:08.177534 1522 omaha_request_action.cc:636] Omaha request network transfer failed. Oct 13 00:05:08.177633 update_engine[1522]: I20251013 00:05:08.177561 1522 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Oct 13 00:05:08.177633 update_engine[1522]: I20251013 00:05:08.177567 1522 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Oct 13 00:05:08.177633 update_engine[1522]: I20251013 00:05:08.177573 1522 update_attempter.cc:306] Processing Done. Oct 13 00:05:08.177633 update_engine[1522]: E20251013 00:05:08.177592 1522 update_attempter.cc:619] Update failed. Oct 13 00:05:08.177633 update_engine[1522]: I20251013 00:05:08.177599 1522 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Oct 13 00:05:08.177633 update_engine[1522]: I20251013 00:05:08.177606 1522 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Oct 13 00:05:08.177633 update_engine[1522]: I20251013 00:05:08.177613 1522 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Oct 13 00:05:08.177851 update_engine[1522]: I20251013 00:05:08.177733 1522 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Oct 13 00:05:08.177851 update_engine[1522]: I20251013 00:05:08.177763 1522 omaha_request_action.cc:271] Posting an Omaha request to disabled Oct 13 00:05:08.177851 update_engine[1522]: I20251013 00:05:08.177770 1522 omaha_request_action.cc:272] Request: Oct 13 00:05:08.177851 update_engine[1522]: Oct 13 00:05:08.177851 update_engine[1522]: Oct 13 00:05:08.177851 update_engine[1522]: Oct 13 00:05:08.177851 update_engine[1522]: Oct 13 00:05:08.177851 update_engine[1522]: Oct 13 00:05:08.177851 update_engine[1522]: Oct 13 00:05:08.177851 update_engine[1522]: I20251013 00:05:08.177778 1522 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Oct 13 00:05:08.177851 update_engine[1522]: I20251013 00:05:08.177804 1522 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Oct 13 00:05:08.178703 locksmithd[1564]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178617 1522 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Oct 13 00:05:08.179025 update_engine[1522]: E20251013 00:05:08.178829 1522 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178900 1522 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178910 1522 omaha_request_action.cc:617] Omaha request response: Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178918 1522 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178925 1522 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178930 1522 update_attempter.cc:306] Processing Done. Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178938 1522 update_attempter.cc:310] Error event sent. Oct 13 00:05:08.179025 update_engine[1522]: I20251013 00:05:08.178947 1522 update_check_scheduler.cc:74] Next update check in 44m29s Oct 13 00:05:08.179463 locksmithd[1564]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Oct 13 00:05:15.686268 containerd[1544]: time="2025-10-13T00:05:15.686174406Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"520b990c7816a623f57d22e23994110ffa535d4f9f9dcc83a4961d14389cf199\" pid:5469 exited_at:{seconds:1760313915 nanos:685390146}" Oct 13 00:05:21.035926 containerd[1544]: time="2025-10-13T00:05:21.035859267Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"83105418c3d5fe7383393bdc3daf1fd43af48494e2dde3ca63fe40d4713ab672\" pid:5492 exited_at:{seconds:1760313921 nanos:35503977}" Oct 13 00:05:26.818252 containerd[1544]: time="2025-10-13T00:05:26.818193213Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"adb9384706089296f0c53781c7489ec611a902cbfbb6816f83dcdfa225068c84\" pid:5514 exited_at:{seconds:1760313926 nanos:817550674}" Oct 13 00:05:31.462851 systemd[1]: Started sshd@7-49.13.69.12:22-139.178.89.65:44400.service - OpenSSH per-connection server daemon (139.178.89.65:44400). Oct 13 00:05:32.462188 sshd[5530]: Accepted publickey for core from 139.178.89.65 port 44400 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:05:32.467060 sshd-session[5530]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:05:32.479404 systemd-logind[1521]: New session 8 of user core. Oct 13 00:05:32.487757 systemd[1]: Started session-8.scope - Session 8 of User core. Oct 13 00:05:33.288591 sshd[5533]: Connection closed by 139.178.89.65 port 44400 Oct 13 00:05:33.290092 sshd-session[5530]: pam_unix(sshd:session): session closed for user core Oct 13 00:05:33.298089 systemd[1]: sshd@7-49.13.69.12:22-139.178.89.65:44400.service: Deactivated successfully. Oct 13 00:05:33.303975 systemd[1]: session-8.scope: Deactivated successfully. Oct 13 00:05:33.307710 systemd-logind[1521]: Session 8 logged out. Waiting for processes to exit. Oct 13 00:05:33.315430 systemd-logind[1521]: Removed session 8. Oct 13 00:05:37.680437 containerd[1544]: time="2025-10-13T00:05:37.680370220Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"a58614b87426aaafe64119321e98a80896eb47f9b169859500d348d1d4027c46\" pid:5558 exited_at:{seconds:1760313937 nanos:679026858}" Oct 13 00:05:38.455714 systemd[1]: Started sshd@8-49.13.69.12:22-139.178.89.65:58282.service - OpenSSH per-connection server daemon (139.178.89.65:58282). Oct 13 00:05:39.459172 sshd[5571]: Accepted publickey for core from 139.178.89.65 port 58282 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:05:39.462787 sshd-session[5571]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:05:39.469887 systemd-logind[1521]: New session 9 of user core. Oct 13 00:05:39.479077 systemd[1]: Started session-9.scope - Session 9 of User core. Oct 13 00:05:40.231467 sshd[5574]: Connection closed by 139.178.89.65 port 58282 Oct 13 00:05:40.232827 sshd-session[5571]: pam_unix(sshd:session): session closed for user core Oct 13 00:05:40.240706 systemd-logind[1521]: Session 9 logged out. Waiting for processes to exit. Oct 13 00:05:40.241493 systemd[1]: sshd@8-49.13.69.12:22-139.178.89.65:58282.service: Deactivated successfully. Oct 13 00:05:40.244479 systemd[1]: session-9.scope: Deactivated successfully. Oct 13 00:05:40.247334 systemd-logind[1521]: Removed session 9. Oct 13 00:05:45.406900 systemd[1]: Started sshd@9-49.13.69.12:22-139.178.89.65:35282.service - OpenSSH per-connection server daemon (139.178.89.65:35282). Oct 13 00:05:45.685499 containerd[1544]: time="2025-10-13T00:05:45.685127168Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"d1e0dd178aba0c733c35902444fff1ec4f371b89759bef4cdf15b8dc028d0893\" pid:5603 exited_at:{seconds:1760313945 nanos:684699754}" Oct 13 00:05:46.417635 sshd[5588]: Accepted publickey for core from 139.178.89.65 port 35282 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:05:46.419954 sshd-session[5588]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:05:46.427873 systemd-logind[1521]: New session 10 of user core. Oct 13 00:05:46.436706 systemd[1]: Started session-10.scope - Session 10 of User core. Oct 13 00:05:47.196018 sshd[5613]: Connection closed by 139.178.89.65 port 35282 Oct 13 00:05:47.197384 sshd-session[5588]: pam_unix(sshd:session): session closed for user core Oct 13 00:05:47.203277 systemd-logind[1521]: Session 10 logged out. Waiting for processes to exit. Oct 13 00:05:47.203834 systemd[1]: sshd@9-49.13.69.12:22-139.178.89.65:35282.service: Deactivated successfully. Oct 13 00:05:47.211465 systemd[1]: session-10.scope: Deactivated successfully. Oct 13 00:05:47.215929 systemd-logind[1521]: Removed session 10. Oct 13 00:05:47.375699 systemd[1]: Started sshd@10-49.13.69.12:22-139.178.89.65:35294.service - OpenSSH per-connection server daemon (139.178.89.65:35294). Oct 13 00:05:48.360000 sshd[5625]: Accepted publickey for core from 139.178.89.65 port 35294 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:05:48.362215 sshd-session[5625]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:05:48.371685 systemd-logind[1521]: New session 11 of user core. Oct 13 00:05:48.377556 systemd[1]: Started session-11.scope - Session 11 of User core. Oct 13 00:05:49.159640 sshd[5628]: Connection closed by 139.178.89.65 port 35294 Oct 13 00:05:49.160712 sshd-session[5625]: pam_unix(sshd:session): session closed for user core Oct 13 00:05:49.165095 systemd-logind[1521]: Session 11 logged out. Waiting for processes to exit. Oct 13 00:05:49.165386 systemd[1]: sshd@10-49.13.69.12:22-139.178.89.65:35294.service: Deactivated successfully. Oct 13 00:05:49.167887 systemd[1]: session-11.scope: Deactivated successfully. Oct 13 00:05:49.170782 systemd-logind[1521]: Removed session 11. Oct 13 00:05:49.326115 systemd[1]: Started sshd@11-49.13.69.12:22-139.178.89.65:35306.service - OpenSSH per-connection server daemon (139.178.89.65:35306). Oct 13 00:05:50.299471 sshd[5637]: Accepted publickey for core from 139.178.89.65 port 35306 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:05:50.301719 sshd-session[5637]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:05:50.309674 systemd-logind[1521]: New session 12 of user core. Oct 13 00:05:50.324731 systemd[1]: Started session-12.scope - Session 12 of User core. Oct 13 00:05:51.039951 sshd[5645]: Connection closed by 139.178.89.65 port 35306 Oct 13 00:05:51.040988 sshd-session[5637]: pam_unix(sshd:session): session closed for user core Oct 13 00:05:51.048053 systemd-logind[1521]: Session 12 logged out. Waiting for processes to exit. Oct 13 00:05:51.048657 systemd[1]: sshd@11-49.13.69.12:22-139.178.89.65:35306.service: Deactivated successfully. Oct 13 00:05:51.051863 systemd[1]: session-12.scope: Deactivated successfully. Oct 13 00:05:51.055105 systemd-logind[1521]: Removed session 12. Oct 13 00:05:51.339830 containerd[1544]: time="2025-10-13T00:05:51.339709644Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"72d39b2652241b8c01eee62dfcce0ef3824b4e7e88c6ee454d176ec22a356c8c\" pid:5668 exited_at:{seconds:1760313951 nanos:339345752}" Oct 13 00:05:56.214475 systemd[1]: Started sshd@12-49.13.69.12:22-139.178.89.65:52050.service - OpenSSH per-connection server daemon (139.178.89.65:52050). Oct 13 00:05:56.813473 containerd[1544]: time="2025-10-13T00:05:56.813199036Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"4f49fed99f129bc3a3d465269822d985988d2e43ca1feb5dc518cec0bc4f2a8e\" pid:5697 exited_at:{seconds:1760313956 nanos:810399784}" Oct 13 00:05:57.205883 sshd[5681]: Accepted publickey for core from 139.178.89.65 port 52050 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:05:57.208354 sshd-session[5681]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:05:57.220168 systemd-logind[1521]: New session 13 of user core. Oct 13 00:05:57.222641 systemd[1]: Started session-13.scope - Session 13 of User core. Oct 13 00:05:57.945449 sshd[5707]: Connection closed by 139.178.89.65 port 52050 Oct 13 00:05:57.944795 sshd-session[5681]: pam_unix(sshd:session): session closed for user core Oct 13 00:05:57.952457 systemd[1]: sshd@12-49.13.69.12:22-139.178.89.65:52050.service: Deactivated successfully. Oct 13 00:05:57.956294 systemd[1]: session-13.scope: Deactivated successfully. Oct 13 00:05:57.958007 systemd-logind[1521]: Session 13 logged out. Waiting for processes to exit. Oct 13 00:05:57.961153 systemd-logind[1521]: Removed session 13. Oct 13 00:06:03.114902 systemd[1]: Started sshd@13-49.13.69.12:22-139.178.89.65:45806.service - OpenSSH per-connection server daemon (139.178.89.65:45806). Oct 13 00:06:04.093954 sshd[5721]: Accepted publickey for core from 139.178.89.65 port 45806 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:04.096508 sshd-session[5721]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:04.102988 systemd-logind[1521]: New session 14 of user core. Oct 13 00:06:04.114114 systemd[1]: Started session-14.scope - Session 14 of User core. Oct 13 00:06:04.879457 sshd[5724]: Connection closed by 139.178.89.65 port 45806 Oct 13 00:06:04.879217 sshd-session[5721]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:04.889465 systemd[1]: sshd@13-49.13.69.12:22-139.178.89.65:45806.service: Deactivated successfully. Oct 13 00:06:04.895555 systemd[1]: session-14.scope: Deactivated successfully. Oct 13 00:06:04.898333 systemd-logind[1521]: Session 14 logged out. Waiting for processes to exit. Oct 13 00:06:04.901835 systemd-logind[1521]: Removed session 14. Oct 13 00:06:07.591341 containerd[1544]: time="2025-10-13T00:06:07.591294981Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"39d1c45596afe2088b83983893fd48d84d4224a93af6b8b5ab49caa23de4bfdd\" pid:5748 exited_at:{seconds:1760313967 nanos:590908269}" Oct 13 00:06:10.056006 systemd[1]: Started sshd@14-49.13.69.12:22-139.178.89.65:45816.service - OpenSSH per-connection server daemon (139.178.89.65:45816). Oct 13 00:06:11.068131 sshd[5760]: Accepted publickey for core from 139.178.89.65 port 45816 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:11.069636 sshd-session[5760]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:11.079340 systemd-logind[1521]: New session 15 of user core. Oct 13 00:06:11.082684 systemd[1]: Started session-15.scope - Session 15 of User core. Oct 13 00:06:11.854917 sshd[5763]: Connection closed by 139.178.89.65 port 45816 Oct 13 00:06:11.854668 sshd-session[5760]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:11.862836 systemd[1]: sshd@14-49.13.69.12:22-139.178.89.65:45816.service: Deactivated successfully. Oct 13 00:06:11.869524 systemd[1]: session-15.scope: Deactivated successfully. Oct 13 00:06:11.871350 systemd-logind[1521]: Session 15 logged out. Waiting for processes to exit. Oct 13 00:06:11.874726 systemd-logind[1521]: Removed session 15. Oct 13 00:06:12.025977 systemd[1]: Started sshd@15-49.13.69.12:22-139.178.89.65:45824.service - OpenSSH per-connection server daemon (139.178.89.65:45824). Oct 13 00:06:12.999401 sshd[5775]: Accepted publickey for core from 139.178.89.65 port 45824 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:13.004105 sshd-session[5775]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:13.014655 systemd-logind[1521]: New session 16 of user core. Oct 13 00:06:13.019738 systemd[1]: Started session-16.scope - Session 16 of User core. Oct 13 00:06:13.918517 sshd[5778]: Connection closed by 139.178.89.65 port 45824 Oct 13 00:06:13.919453 sshd-session[5775]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:13.925516 systemd-logind[1521]: Session 16 logged out. Waiting for processes to exit. Oct 13 00:06:13.926007 systemd[1]: sshd@15-49.13.69.12:22-139.178.89.65:45824.service: Deactivated successfully. Oct 13 00:06:13.931357 systemd[1]: session-16.scope: Deactivated successfully. Oct 13 00:06:13.936264 systemd-logind[1521]: Removed session 16. Oct 13 00:06:14.089861 systemd[1]: Started sshd@16-49.13.69.12:22-139.178.89.65:55114.service - OpenSSH per-connection server daemon (139.178.89.65:55114). Oct 13 00:06:15.082609 sshd[5788]: Accepted publickey for core from 139.178.89.65 port 55114 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:15.084910 sshd-session[5788]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:15.091752 systemd-logind[1521]: New session 17 of user core. Oct 13 00:06:15.103697 systemd[1]: Started session-17.scope - Session 17 of User core. Oct 13 00:06:15.701595 containerd[1544]: time="2025-10-13T00:06:15.701526440Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"ad1302c05a5f89ce7d010daf2038d8f3ac7dc4c97cec01a604e6095f70944f7d\" pid:5809 exited_at:{seconds:1760313975 nanos:699868307}" Oct 13 00:06:16.573542 sshd[5791]: Connection closed by 139.178.89.65 port 55114 Oct 13 00:06:16.575248 sshd-session[5788]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:16.585730 systemd[1]: sshd@16-49.13.69.12:22-139.178.89.65:55114.service: Deactivated successfully. Oct 13 00:06:16.590446 systemd[1]: session-17.scope: Deactivated successfully. Oct 13 00:06:16.596674 systemd-logind[1521]: Session 17 logged out. Waiting for processes to exit. Oct 13 00:06:16.601833 systemd-logind[1521]: Removed session 17. Oct 13 00:06:16.748028 systemd[1]: Started sshd@17-49.13.69.12:22-139.178.89.65:55120.service - OpenSSH per-connection server daemon (139.178.89.65:55120). Oct 13 00:06:17.753624 sshd[5831]: Accepted publickey for core from 139.178.89.65 port 55120 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:17.756032 sshd-session[5831]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:17.767462 systemd-logind[1521]: New session 18 of user core. Oct 13 00:06:17.773386 systemd[1]: Started session-18.scope - Session 18 of User core. Oct 13 00:06:18.643572 sshd[5834]: Connection closed by 139.178.89.65 port 55120 Oct 13 00:06:18.644274 sshd-session[5831]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:18.651506 systemd[1]: sshd@17-49.13.69.12:22-139.178.89.65:55120.service: Deactivated successfully. Oct 13 00:06:18.655033 systemd[1]: session-18.scope: Deactivated successfully. Oct 13 00:06:18.658515 systemd-logind[1521]: Session 18 logged out. Waiting for processes to exit. Oct 13 00:06:18.661881 systemd-logind[1521]: Removed session 18. Oct 13 00:06:18.816095 systemd[1]: Started sshd@18-49.13.69.12:22-139.178.89.65:55134.service - OpenSSH per-connection server daemon (139.178.89.65:55134). Oct 13 00:06:19.806011 sshd[5843]: Accepted publickey for core from 139.178.89.65 port 55134 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:19.806803 sshd-session[5843]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:19.815098 systemd-logind[1521]: New session 19 of user core. Oct 13 00:06:19.821658 systemd[1]: Started session-19.scope - Session 19 of User core. Oct 13 00:06:20.551619 sshd[5846]: Connection closed by 139.178.89.65 port 55134 Oct 13 00:06:20.551338 sshd-session[5843]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:20.557340 systemd[1]: sshd@18-49.13.69.12:22-139.178.89.65:55134.service: Deactivated successfully. Oct 13 00:06:20.560378 systemd[1]: session-19.scope: Deactivated successfully. Oct 13 00:06:20.562927 systemd-logind[1521]: Session 19 logged out. Waiting for processes to exit. Oct 13 00:06:20.565272 systemd-logind[1521]: Removed session 19. Oct 13 00:06:21.020872 containerd[1544]: time="2025-10-13T00:06:21.020831885Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"27956be3053c510efe580b3d337b7cda6c5bdf7ed38a3ecf97e7b353d6aa0f00\" pid:5878 exited_at:{seconds:1760313981 nanos:20441290}" Oct 13 00:06:25.715796 systemd[1]: Started sshd@19-49.13.69.12:22-139.178.89.65:57410.service - OpenSSH per-connection server daemon (139.178.89.65:57410). Oct 13 00:06:26.692606 sshd[5892]: Accepted publickey for core from 139.178.89.65 port 57410 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:26.694636 sshd-session[5892]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:26.701671 systemd-logind[1521]: New session 20 of user core. Oct 13 00:06:26.707700 systemd[1]: Started session-20.scope - Session 20 of User core. Oct 13 00:06:26.848340 containerd[1544]: time="2025-10-13T00:06:26.848192776Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"804920c08f4259a163eb42a02756ba6821a26d831fb13faea73fb5fd9b1d3d7a\" pid:5909 exited_at:{seconds:1760313986 nanos:847482863}" Oct 13 00:06:27.458427 sshd[5895]: Connection closed by 139.178.89.65 port 57410 Oct 13 00:06:27.459630 sshd-session[5892]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:27.467999 systemd[1]: sshd@19-49.13.69.12:22-139.178.89.65:57410.service: Deactivated successfully. Oct 13 00:06:27.468054 systemd-logind[1521]: Session 20 logged out. Waiting for processes to exit. Oct 13 00:06:27.474047 systemd[1]: session-20.scope: Deactivated successfully. Oct 13 00:06:27.478556 systemd-logind[1521]: Removed session 20. Oct 13 00:06:30.164801 systemd[1]: Started sshd@20-49.13.69.12:22-59.94.176.55:60508.service - OpenSSH per-connection server daemon (59.94.176.55:60508). Oct 13 00:06:30.547297 sshd[5931]: Connection closed by 59.94.176.55 port 60508 [preauth] Oct 13 00:06:30.551549 systemd[1]: sshd@20-49.13.69.12:22-59.94.176.55:60508.service: Deactivated successfully. Oct 13 00:06:32.633499 systemd[1]: Started sshd@21-49.13.69.12:22-139.178.89.65:39228.service - OpenSSH per-connection server daemon (139.178.89.65:39228). Oct 13 00:06:33.621052 sshd[5939]: Accepted publickey for core from 139.178.89.65 port 39228 ssh2: RSA SHA256:9hygYNV3qUvJXdEWXIOx7wVbZ1g8nxoR791t75pOQbs Oct 13 00:06:33.625203 sshd-session[5939]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 00:06:33.635088 systemd-logind[1521]: New session 21 of user core. Oct 13 00:06:33.639690 systemd[1]: Started session-21.scope - Session 21 of User core. Oct 13 00:06:34.370518 sshd[5942]: Connection closed by 139.178.89.65 port 39228 Oct 13 00:06:34.371636 sshd-session[5939]: pam_unix(sshd:session): session closed for user core Oct 13 00:06:34.379385 systemd-logind[1521]: Session 21 logged out. Waiting for processes to exit. Oct 13 00:06:34.380179 systemd[1]: sshd@21-49.13.69.12:22-139.178.89.65:39228.service: Deactivated successfully. Oct 13 00:06:34.385146 systemd[1]: session-21.scope: Deactivated successfully. Oct 13 00:06:34.391823 systemd-logind[1521]: Removed session 21. Oct 13 00:06:37.632385 containerd[1544]: time="2025-10-13T00:06:37.632320564Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"e08ae6b008f80f2ca596d57ee71622b109786e3b381a9205460522372654c00c\" pid:5965 exited_at:{seconds:1760313997 nanos:631953485}" Oct 13 00:06:45.684288 containerd[1544]: time="2025-10-13T00:06:45.684154839Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"25c16e13b9803a014b72631cd12e388534b431c70000707787c67b66b634e953\" pid:5997 exited_at:{seconds:1760314005 nanos:683854079}" Oct 13 00:06:51.336366 containerd[1544]: time="2025-10-13T00:06:51.336222250Z" level=info msg="TaskExit event in podsandbox handler container_id:\"143ac39307c705d2e5ec129e0aeffb90eb7f2a3c35d7640d9ff3b4eb06eb3184\" id:\"e093ef33bd811dd7f63715b88c96b767811a213690c699363ade29aee6ad94de\" pid:6031 exited_at:{seconds:1760314011 nanos:335044206}" Oct 13 00:06:56.815959 containerd[1544]: time="2025-10-13T00:06:56.815903919Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e180e3e4c8c1ae6f55ae7fb16aadd5112bb8c8f0a834d4a909ed4ea13156bace\" id:\"de0d620d3d1971c399dad71046076c253ca27fffb60d5e5ea56c59c2f63b3c97\" pid:6055 exited_at:{seconds:1760314016 nanos:814663593}" Oct 13 00:07:07.069908 systemd[1]: cri-containerd-5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7.scope: Deactivated successfully. Oct 13 00:07:07.072160 systemd[1]: cri-containerd-5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7.scope: Consumed 23.842s CPU time, 109.8M memory peak, 4.7M read from disk. Oct 13 00:07:07.075081 containerd[1544]: time="2025-10-13T00:07:07.074807360Z" level=info msg="received exit event container_id:\"5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7\" id:\"5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7\" pid:3053 exit_status:1 exited_at:{seconds:1760314027 nanos:74346916}" Oct 13 00:07:07.075767 containerd[1544]: time="2025-10-13T00:07:07.075121482Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7\" id:\"5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7\" pid:3053 exit_status:1 exited_at:{seconds:1760314027 nanos:74346916}" Oct 13 00:07:07.108867 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7-rootfs.mount: Deactivated successfully. Oct 13 00:07:07.204267 kubelet[2740]: E1013 00:07:07.203774 2740 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:43574->10.0.0.2:2379: read: connection timed out" Oct 13 00:07:07.207482 systemd[1]: cri-containerd-cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798.scope: Deactivated successfully. Oct 13 00:07:07.207975 systemd[1]: cri-containerd-cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798.scope: Consumed 4.830s CPU time, 27.3M memory peak, 3.5M read from disk. Oct 13 00:07:07.213399 containerd[1544]: time="2025-10-13T00:07:07.213305823Z" level=info msg="received exit event container_id:\"cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798\" id:\"cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798\" pid:2593 exit_status:1 exited_at:{seconds:1760314027 nanos:212961540}" Oct 13 00:07:07.213846 containerd[1544]: time="2025-10-13T00:07:07.213573506Z" level=info msg="TaskExit event in podsandbox handler container_id:\"cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798\" id:\"cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798\" pid:2593 exit_status:1 exited_at:{seconds:1760314027 nanos:212961540}" Oct 13 00:07:07.246320 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798-rootfs.mount: Deactivated successfully. Oct 13 00:07:07.432802 kubelet[2740]: I1013 00:07:07.432755 2740 scope.go:117] "RemoveContainer" containerID="5dfa2405d7f4d7333737e85cc1e87b0711627befbae323561ff1b48ec570bac7" Oct 13 00:07:07.434781 kubelet[2740]: I1013 00:07:07.434636 2740 scope.go:117] "RemoveContainer" containerID="cccaf2024ce787f1dc4b91575e679cf02bdb656308ee940a37a0ff926807a798" Oct 13 00:07:07.437947 containerd[1544]: time="2025-10-13T00:07:07.437898783Z" level=info msg="CreateContainer within sandbox \"878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Oct 13 00:07:07.438441 containerd[1544]: time="2025-10-13T00:07:07.438242706Z" level=info msg="CreateContainer within sandbox \"4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Oct 13 00:07:07.457194 containerd[1544]: time="2025-10-13T00:07:07.455905697Z" level=info msg="Container 26a4112d57b6315cfe679d5c254eb41ca5cb9f008c350b1e0996000a02bbf202: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:07:07.463537 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4027960871.mount: Deactivated successfully. Oct 13 00:07:07.470738 containerd[1544]: time="2025-10-13T00:07:07.468840407Z" level=info msg="Container c4b193c72fb89c187ad9a65382549346a7cd29874ad8b7db4368bf71b8537e40: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:07:07.486888 containerd[1544]: time="2025-10-13T00:07:07.486814681Z" level=info msg="CreateContainer within sandbox \"4b983c266e8dd7be3cb7bc7fbd230e3d8135208f18e7e19f96517146f54628c1\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"26a4112d57b6315cfe679d5c254eb41ca5cb9f008c350b1e0996000a02bbf202\"" Oct 13 00:07:07.494731 containerd[1544]: time="2025-10-13T00:07:07.494656828Z" level=info msg="CreateContainer within sandbox \"878166740480b9930de509866f17c8f9fd893fce0f01fdba608f0921508c6982\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"c4b193c72fb89c187ad9a65382549346a7cd29874ad8b7db4368bf71b8537e40\"" Oct 13 00:07:07.496920 systemd[1]: cri-containerd-e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5.scope: Deactivated successfully. Oct 13 00:07:07.497255 systemd[1]: cri-containerd-e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5.scope: Consumed 6.235s CPU time, 68.5M memory peak, 3.5M read from disk. Oct 13 00:07:07.505804 containerd[1544]: time="2025-10-13T00:07:07.505625682Z" level=info msg="StartContainer for \"26a4112d57b6315cfe679d5c254eb41ca5cb9f008c350b1e0996000a02bbf202\"" Oct 13 00:07:07.509683 containerd[1544]: time="2025-10-13T00:07:07.509532395Z" level=info msg="StartContainer for \"c4b193c72fb89c187ad9a65382549346a7cd29874ad8b7db4368bf71b8537e40\"" Oct 13 00:07:07.513011 containerd[1544]: time="2025-10-13T00:07:07.511145089Z" level=info msg="connecting to shim c4b193c72fb89c187ad9a65382549346a7cd29874ad8b7db4368bf71b8537e40" address="unix:///run/containerd/s/b1102b39bd6fa2999ccdc54d7af04c1e31b62d5f2bb1b27426ed52f81abd3d6f" protocol=ttrpc version=3 Oct 13 00:07:07.513011 containerd[1544]: time="2025-10-13T00:07:07.511569373Z" level=info msg="connecting to shim 26a4112d57b6315cfe679d5c254eb41ca5cb9f008c350b1e0996000a02bbf202" address="unix:///run/containerd/s/8acae749befb3ed7270aa02c8eb5a978b36c1cc07ff988806a9b33f48a499b3d" protocol=ttrpc version=3 Oct 13 00:07:07.513011 containerd[1544]: time="2025-10-13T00:07:07.512874104Z" level=info msg="received exit event container_id:\"e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5\" id:\"e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5\" pid:2561 exit_status:1 exited_at:{seconds:1760314027 nanos:510536804}" Oct 13 00:07:07.515159 containerd[1544]: time="2025-10-13T00:07:07.514997842Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5\" id:\"e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5\" pid:2561 exit_status:1 exited_at:{seconds:1760314027 nanos:510536804}" Oct 13 00:07:07.560696 systemd[1]: Started cri-containerd-26a4112d57b6315cfe679d5c254eb41ca5cb9f008c350b1e0996000a02bbf202.scope - libcontainer container 26a4112d57b6315cfe679d5c254eb41ca5cb9f008c350b1e0996000a02bbf202. Oct 13 00:07:07.563289 systemd[1]: Started cri-containerd-c4b193c72fb89c187ad9a65382549346a7cd29874ad8b7db4368bf71b8537e40.scope - libcontainer container c4b193c72fb89c187ad9a65382549346a7cd29874ad8b7db4368bf71b8537e40. Oct 13 00:07:07.660511 containerd[1544]: time="2025-10-13T00:07:07.660218043Z" level=info msg="StartContainer for \"c4b193c72fb89c187ad9a65382549346a7cd29874ad8b7db4368bf71b8537e40\" returns successfully" Oct 13 00:07:07.662091 containerd[1544]: time="2025-10-13T00:07:07.661185411Z" level=info msg="StartContainer for \"26a4112d57b6315cfe679d5c254eb41ca5cb9f008c350b1e0996000a02bbf202\" returns successfully" Oct 13 00:07:07.765172 containerd[1544]: time="2025-10-13T00:07:07.765011979Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2200529f186acb9d0ad30fefbf45373b12bc53f99f5630b7b56c85cc9e331e02\" id:\"a09746b96cfb8d91ffa1bb4b891aa6d77143adb7bcfe6b9cadebc535fce37408\" pid:6137 exited_at:{seconds:1760314027 nanos:764638655}" Oct 13 00:07:08.109220 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5-rootfs.mount: Deactivated successfully. Oct 13 00:07:08.442532 kubelet[2740]: I1013 00:07:08.442362 2740 scope.go:117] "RemoveContainer" containerID="e8b63f997e196713bde6daca9bad9f412a9ad9b97042c1d8b775bb8014fefae5" Oct 13 00:07:08.449272 containerd[1544]: time="2025-10-13T00:07:08.449203402Z" level=info msg="CreateContainer within sandbox \"67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Oct 13 00:07:08.476182 containerd[1544]: time="2025-10-13T00:07:08.475625996Z" level=info msg="Container 9e1056aa2df1d429ed836c521771cacdd0a42ee5768583a5932a4a962ce82644: CDI devices from CRI Config.CDIDevices: []" Oct 13 00:07:08.490246 containerd[1544]: time="2025-10-13T00:07:08.490104364Z" level=info msg="CreateContainer within sandbox \"67f7756c9b639062307a667ec29cc0319fba3ea5d3c31e145a3e50d41e13a226\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"9e1056aa2df1d429ed836c521771cacdd0a42ee5768583a5932a4a962ce82644\"" Oct 13 00:07:08.491023 containerd[1544]: time="2025-10-13T00:07:08.490954332Z" level=info msg="StartContainer for \"9e1056aa2df1d429ed836c521771cacdd0a42ee5768583a5932a4a962ce82644\"" Oct 13 00:07:08.492795 containerd[1544]: time="2025-10-13T00:07:08.492721627Z" level=info msg="connecting to shim 9e1056aa2df1d429ed836c521771cacdd0a42ee5768583a5932a4a962ce82644" address="unix:///run/containerd/s/89567c988c35bb5f53ce13c330eeefbd9f42c9d384d07e4e10c9be79a3a0f0f7" protocol=ttrpc version=3 Oct 13 00:07:08.526896 systemd[1]: Started cri-containerd-9e1056aa2df1d429ed836c521771cacdd0a42ee5768583a5932a4a962ce82644.scope - libcontainer container 9e1056aa2df1d429ed836c521771cacdd0a42ee5768583a5932a4a962ce82644. Oct 13 00:07:08.591751 containerd[1544]: time="2025-10-13T00:07:08.591435501Z" level=info msg="StartContainer for \"9e1056aa2df1d429ed836c521771cacdd0a42ee5768583a5932a4a962ce82644\" returns successfully" Oct 13 00:07:09.576563 kubelet[2740]: E1013 00:07:09.569521 2740 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:43410->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4459-1-0-3-bea61e6062.186de446183bfdde kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4459-1-0-3-bea61e6062,UID:cb2952b2b29a7f47b55f00542af22ebf,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4459-1-0-3-bea61e6062,},FirstTimestamp:2025-10-13 00:06:59.077160414 +0000 UTC m=+246.004023753,LastTimestamp:2025-10-13 00:06:59.077160414 +0000 UTC m=+246.004023753,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459-1-0-3-bea61e6062,}"