Apr 23 23:18:02.049740 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd490] Apr 23 23:18:02.049757 kernel: Linux version 6.12.81-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Thu Apr 23 21:57:58 -00 2026 Apr 23 23:18:02.049763 kernel: KASLR enabled Apr 23 23:18:02.049767 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Apr 23 23:18:02.049771 kernel: printk: legacy bootconsole [pl11] enabled Apr 23 23:18:02.049776 kernel: efi: EFI v2.7 by EDK II Apr 23 23:18:02.049781 kernel: efi: ACPI 2.0=0x3f979018 SMBIOS=0x3f8a0000 SMBIOS 3.0=0x3f880000 MEMATTR=0x3e3f8018 RNG=0x3f979998 MEMRESERVE=0x3db83598 Apr 23 23:18:02.049785 kernel: random: crng init done Apr 23 23:18:02.049789 kernel: secureboot: Secure boot disabled Apr 23 23:18:02.049793 kernel: ACPI: Early table checksum verification disabled Apr 23 23:18:02.049797 kernel: ACPI: RSDP 0x000000003F979018 000024 (v02 VRTUAL) Apr 23 23:18:02.049801 kernel: ACPI: XSDT 0x000000003F979F18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049805 kernel: ACPI: FACP 0x000000003F979C18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049809 kernel: ACPI: DSDT 0x000000003F95A018 01E046 (v02 MSFTVM DSDT01 00000001 INTL 20230628) Apr 23 23:18:02.049815 kernel: ACPI: DBG2 0x000000003F979B18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049829 kernel: ACPI: GTDT 0x000000003F979D98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049834 kernel: ACPI: OEM0 0x000000003F979098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049838 kernel: ACPI: SPCR 0x000000003F979A98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049842 kernel: ACPI: APIC 0x000000003F979818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049848 kernel: ACPI: SRAT 0x000000003F979198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049852 kernel: ACPI: PPTT 0x000000003F979418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Apr 23 23:18:02.049856 kernel: ACPI: BGRT 0x000000003F979E98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 23 23:18:02.049860 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Apr 23 23:18:02.049865 kernel: ACPI: Use ACPI SPCR as default console: Yes Apr 23 23:18:02.049869 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Apr 23 23:18:02.049873 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] hotplug Apr 23 23:18:02.049877 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] hotplug Apr 23 23:18:02.049881 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Apr 23 23:18:02.049886 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Apr 23 23:18:02.049890 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Apr 23 23:18:02.049895 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Apr 23 23:18:02.049899 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Apr 23 23:18:02.049903 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Apr 23 23:18:02.049907 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Apr 23 23:18:02.049911 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Apr 23 23:18:02.049916 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Apr 23 23:18:02.049920 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x1bfffffff] -> [mem 0x00000000-0x1bfffffff] Apr 23 23:18:02.049924 kernel: NODE_DATA(0) allocated [mem 0x1bf7ffa00-0x1bf806fff] Apr 23 23:18:02.049928 kernel: Zone ranges: Apr 23 23:18:02.049932 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Apr 23 23:18:02.049939 kernel: DMA32 empty Apr 23 23:18:02.049943 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Apr 23 23:18:02.049948 kernel: Device empty Apr 23 23:18:02.049952 kernel: Movable zone start for each node Apr 23 23:18:02.049956 kernel: Early memory node ranges Apr 23 23:18:02.049961 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Apr 23 23:18:02.049966 kernel: node 0: [mem 0x0000000000824000-0x000000003f38ffff] Apr 23 23:18:02.049970 kernel: node 0: [mem 0x000000003f390000-0x000000003f93ffff] Apr 23 23:18:02.049975 kernel: node 0: [mem 0x000000003f940000-0x000000003f9effff] Apr 23 23:18:02.049979 kernel: node 0: [mem 0x000000003f9f0000-0x000000003fdeffff] Apr 23 23:18:02.049983 kernel: node 0: [mem 0x000000003fdf0000-0x000000003fffffff] Apr 23 23:18:02.049987 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Apr 23 23:18:02.049992 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Apr 23 23:18:02.049996 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Apr 23 23:18:02.050000 kernel: cma: Reserved 16 MiB at 0x000000003ca00000 on node -1 Apr 23 23:18:02.050005 kernel: psci: probing for conduit method from ACPI. Apr 23 23:18:02.050009 kernel: psci: PSCIv1.3 detected in firmware. Apr 23 23:18:02.050013 kernel: psci: Using standard PSCI v0.2 function IDs Apr 23 23:18:02.050018 kernel: psci: MIGRATE_INFO_TYPE not supported. Apr 23 23:18:02.050023 kernel: psci: SMC Calling Convention v1.4 Apr 23 23:18:02.050027 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Apr 23 23:18:02.050031 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Apr 23 23:18:02.050036 kernel: percpu: Embedded 33 pages/cpu s97752 r8192 d29224 u135168 Apr 23 23:18:02.050040 kernel: pcpu-alloc: s97752 r8192 d29224 u135168 alloc=33*4096 Apr 23 23:18:02.050045 kernel: pcpu-alloc: [0] 0 [0] 1 Apr 23 23:18:02.050049 kernel: Detected PIPT I-cache on CPU0 Apr 23 23:18:02.050053 kernel: CPU features: detected: Address authentication (architected QARMA5 algorithm) Apr 23 23:18:02.050058 kernel: CPU features: detected: GIC system register CPU interface Apr 23 23:18:02.050062 kernel: CPU features: detected: Spectre-v4 Apr 23 23:18:02.050066 kernel: CPU features: detected: Spectre-BHB Apr 23 23:18:02.050071 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 23 23:18:02.050076 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 23 23:18:02.050080 kernel: CPU features: detected: ARM erratum 2067961 or 2054223 Apr 23 23:18:02.050084 kernel: CPU features: detected: SSBS not fully self-synchronizing Apr 23 23:18:02.050088 kernel: alternatives: applying boot alternatives Apr 23 23:18:02.050094 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=8669c84e6bfac0c003f3ced682d9b5c0fda27fc2948639441be65941607b4c3d Apr 23 23:18:02.050098 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 23 23:18:02.050103 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 23 23:18:02.050107 kernel: Fallback order for Node 0: 0 Apr 23 23:18:02.050111 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1048540 Apr 23 23:18:02.050117 kernel: Policy zone: Normal Apr 23 23:18:02.050121 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 23 23:18:02.050125 kernel: software IO TLB: area num 2. Apr 23 23:18:02.050130 kernel: software IO TLB: mapped [mem 0x00000000358f0000-0x00000000398f0000] (64MB) Apr 23 23:18:02.050134 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 23 23:18:02.050138 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 23 23:18:02.050143 kernel: rcu: RCU event tracing is enabled. Apr 23 23:18:02.050148 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 23 23:18:02.050152 kernel: Trampoline variant of Tasks RCU enabled. Apr 23 23:18:02.050156 kernel: Tracing variant of Tasks RCU enabled. Apr 23 23:18:02.050161 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 23 23:18:02.050165 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 23 23:18:02.050170 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 23 23:18:02.050175 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 23 23:18:02.050179 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 23 23:18:02.050183 kernel: GICv3: 960 SPIs implemented Apr 23 23:18:02.050188 kernel: GICv3: 0 Extended SPIs implemented Apr 23 23:18:02.050192 kernel: Root IRQ handler: gic_handle_irq Apr 23 23:18:02.050196 kernel: GICv3: GICv3 features: 16 PPIs, RSS Apr 23 23:18:02.050201 kernel: GICv3: GICD_CTRL.DS=0, SCR_EL3.FIQ=0 Apr 23 23:18:02.050205 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Apr 23 23:18:02.050209 kernel: ITS: No ITS available, not enabling LPIs Apr 23 23:18:02.050214 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 23 23:18:02.050219 kernel: arch_timer: cp15 timer(s) running at 1000.00MHz (virt). Apr 23 23:18:02.050223 kernel: clocksource: arch_sys_counter: mask: 0x1fffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Apr 23 23:18:02.050228 kernel: sched_clock: 61 bits at 1000MHz, resolution 1ns, wraps every 4398046511103ns Apr 23 23:18:02.050232 kernel: Console: colour dummy device 80x25 Apr 23 23:18:02.050237 kernel: printk: legacy console [tty1] enabled Apr 23 23:18:02.050241 kernel: ACPI: Core revision 20240827 Apr 23 23:18:02.050246 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 2000.00 BogoMIPS (lpj=1000000) Apr 23 23:18:02.050251 kernel: pid_max: default: 32768 minimum: 301 Apr 23 23:18:02.050255 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Apr 23 23:18:02.050260 kernel: landlock: Up and running. Apr 23 23:18:02.050265 kernel: SELinux: Initializing. Apr 23 23:18:02.050269 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 23 23:18:02.050274 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 23 23:18:02.050278 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0xa0000e, misc 0x31e1 Apr 23 23:18:02.050283 kernel: Hyper-V: Host Build 10.0.26102.1283-1-0 Apr 23 23:18:02.050290 kernel: Hyper-V: enabling crash_kexec_post_notifiers Apr 23 23:18:02.050296 kernel: rcu: Hierarchical SRCU implementation. Apr 23 23:18:02.050301 kernel: rcu: Max phase no-delay instances is 400. Apr 23 23:18:02.050305 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Apr 23 23:18:02.050310 kernel: Remapping and enabling EFI services. Apr 23 23:18:02.050315 kernel: smp: Bringing up secondary CPUs ... Apr 23 23:18:02.050319 kernel: Detected PIPT I-cache on CPU1 Apr 23 23:18:02.050325 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Apr 23 23:18:02.050330 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd490] Apr 23 23:18:02.050334 kernel: smp: Brought up 1 node, 2 CPUs Apr 23 23:18:02.050339 kernel: SMP: Total of 2 processors activated. Apr 23 23:18:02.050344 kernel: CPU: All CPU(s) started at EL1 Apr 23 23:18:02.050349 kernel: CPU features: detected: 32-bit EL0 Support Apr 23 23:18:02.050354 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Apr 23 23:18:02.050359 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 23 23:18:02.050363 kernel: CPU features: detected: Common not Private translations Apr 23 23:18:02.050368 kernel: CPU features: detected: CRC32 instructions Apr 23 23:18:02.050373 kernel: CPU features: detected: Generic authentication (architected QARMA5 algorithm) Apr 23 23:18:02.050378 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 23 23:18:02.050382 kernel: CPU features: detected: LSE atomic instructions Apr 23 23:18:02.050387 kernel: CPU features: detected: Privileged Access Never Apr 23 23:18:02.050393 kernel: CPU features: detected: Speculation barrier (SB) Apr 23 23:18:02.050398 kernel: CPU features: detected: TLB range maintenance instructions Apr 23 23:18:02.050402 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Apr 23 23:18:02.050407 kernel: CPU features: detected: Scalable Vector Extension Apr 23 23:18:02.050412 kernel: alternatives: applying system-wide alternatives Apr 23 23:18:02.050416 kernel: CPU features: detected: Hardware dirty bit management on CPU0-1 Apr 23 23:18:02.050421 kernel: SVE: maximum available vector length 16 bytes per vector Apr 23 23:18:02.050426 kernel: SVE: default vector length 16 bytes per vector Apr 23 23:18:02.050431 kernel: Memory: 3952756K/4194160K available (11200K kernel code, 2458K rwdata, 9092K rodata, 39552K init, 1038K bss, 220208K reserved, 16384K cma-reserved) Apr 23 23:18:02.050436 kernel: devtmpfs: initialized Apr 23 23:18:02.050441 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 23 23:18:02.050446 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 23 23:18:02.050451 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Apr 23 23:18:02.050455 kernel: 0 pages in range for non-PLT usage Apr 23 23:18:02.050460 kernel: 508384 pages in range for PLT usage Apr 23 23:18:02.050465 kernel: pinctrl core: initialized pinctrl subsystem Apr 23 23:18:02.050469 kernel: SMBIOS 3.1.0 present. Apr 23 23:18:02.050475 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 01/08/2026 Apr 23 23:18:02.050480 kernel: DMI: Memory slots populated: 2/2 Apr 23 23:18:02.050484 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 23 23:18:02.050489 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Apr 23 23:18:02.050494 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 23 23:18:02.050499 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 23 23:18:02.050503 kernel: audit: initializing netlink subsys (disabled) Apr 23 23:18:02.050508 kernel: audit: type=2000 audit(0.059:1): state=initialized audit_enabled=0 res=1 Apr 23 23:18:02.050513 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 23 23:18:02.050518 kernel: cpuidle: using governor menu Apr 23 23:18:02.050523 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 23 23:18:02.050528 kernel: ASID allocator initialised with 32768 entries Apr 23 23:18:02.050532 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 23 23:18:02.050537 kernel: Serial: AMBA PL011 UART driver Apr 23 23:18:02.050542 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 23 23:18:02.050546 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Apr 23 23:18:02.050551 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Apr 23 23:18:02.050556 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Apr 23 23:18:02.050561 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 23 23:18:02.050566 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Apr 23 23:18:02.050571 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Apr 23 23:18:02.050575 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Apr 23 23:18:02.050580 kernel: ACPI: Added _OSI(Module Device) Apr 23 23:18:02.050585 kernel: ACPI: Added _OSI(Processor Device) Apr 23 23:18:02.050589 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 23 23:18:02.050594 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 23 23:18:02.050599 kernel: ACPI: Interpreter enabled Apr 23 23:18:02.050604 kernel: ACPI: Using GIC for interrupt routing Apr 23 23:18:02.050609 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Apr 23 23:18:02.050614 kernel: printk: legacy console [ttyAMA0] enabled Apr 23 23:18:02.050618 kernel: printk: legacy bootconsole [pl11] disabled Apr 23 23:18:02.050623 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Apr 23 23:18:02.050628 kernel: ACPI: CPU0 has been hot-added Apr 23 23:18:02.050633 kernel: ACPI: CPU1 has been hot-added Apr 23 23:18:02.050637 kernel: iommu: Default domain type: Translated Apr 23 23:18:02.050642 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 23 23:18:02.050647 kernel: efivars: Registered efivars operations Apr 23 23:18:02.050652 kernel: vgaarb: loaded Apr 23 23:18:02.050657 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 23 23:18:02.050661 kernel: VFS: Disk quotas dquot_6.6.0 Apr 23 23:18:02.050666 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 23 23:18:02.050671 kernel: pnp: PnP ACPI init Apr 23 23:18:02.050675 kernel: pnp: PnP ACPI: found 0 devices Apr 23 23:18:02.050680 kernel: NET: Registered PF_INET protocol family Apr 23 23:18:02.050685 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 23 23:18:02.050690 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 23 23:18:02.050695 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 23 23:18:02.050700 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 23 23:18:02.050705 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 23 23:18:02.050710 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 23 23:18:02.050714 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 23 23:18:02.050719 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 23 23:18:02.050724 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 23 23:18:02.050728 kernel: PCI: CLS 0 bytes, default 64 Apr 23 23:18:02.050733 kernel: kvm [1]: HYP mode not available Apr 23 23:18:02.050738 kernel: Initialise system trusted keyrings Apr 23 23:18:02.050743 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 23 23:18:02.050748 kernel: Key type asymmetric registered Apr 23 23:18:02.050753 kernel: Asymmetric key parser 'x509' registered Apr 23 23:18:02.050757 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Apr 23 23:18:02.050762 kernel: io scheduler mq-deadline registered Apr 23 23:18:02.050767 kernel: io scheduler kyber registered Apr 23 23:18:02.050771 kernel: io scheduler bfq registered Apr 23 23:18:02.050776 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 23 23:18:02.050782 kernel: thunder_xcv, ver 1.0 Apr 23 23:18:02.050786 kernel: thunder_bgx, ver 1.0 Apr 23 23:18:02.050791 kernel: nicpf, ver 1.0 Apr 23 23:18:02.050796 kernel: nicvf, ver 1.0 Apr 23 23:18:02.050898 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 23 23:18:02.050948 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-04-23T23:18:01 UTC (1776986281) Apr 23 23:18:02.050954 kernel: efifb: probing for efifb Apr 23 23:18:02.050961 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Apr 23 23:18:02.050966 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Apr 23 23:18:02.050970 kernel: efifb: scrolling: redraw Apr 23 23:18:02.050975 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 23 23:18:02.050980 kernel: Console: switching to colour frame buffer device 128x48 Apr 23 23:18:02.050985 kernel: fb0: EFI VGA frame buffer device Apr 23 23:18:02.050989 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Apr 23 23:18:02.050994 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 23 23:18:02.050999 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Apr 23 23:18:02.051004 kernel: watchdog: NMI not fully supported Apr 23 23:18:02.051009 kernel: watchdog: Hard watchdog permanently disabled Apr 23 23:18:02.051014 kernel: NET: Registered PF_INET6 protocol family Apr 23 23:18:02.051018 kernel: Segment Routing with IPv6 Apr 23 23:18:02.051023 kernel: In-situ OAM (IOAM) with IPv6 Apr 23 23:18:02.051028 kernel: NET: Registered PF_PACKET protocol family Apr 23 23:18:02.051033 kernel: Key type dns_resolver registered Apr 23 23:18:02.051037 kernel: registered taskstats version 1 Apr 23 23:18:02.051042 kernel: Loading compiled-in X.509 certificates Apr 23 23:18:02.051047 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.81-flatcar: 1129832e4b4ea3c9ff0dc43e02ec7de2e4d9d907' Apr 23 23:18:02.051052 kernel: Demotion targets for Node 0: null Apr 23 23:18:02.051057 kernel: Key type .fscrypt registered Apr 23 23:18:02.051061 kernel: Key type fscrypt-provisioning registered Apr 23 23:18:02.051066 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 23 23:18:02.051071 kernel: ima: Allocated hash algorithm: sha1 Apr 23 23:18:02.051075 kernel: ima: No architecture policies found Apr 23 23:18:02.051080 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 23 23:18:02.051085 kernel: clk: Disabling unused clocks Apr 23 23:18:02.051089 kernel: PM: genpd: Disabling unused power domains Apr 23 23:18:02.051095 kernel: Warning: unable to open an initial console. Apr 23 23:18:02.051100 kernel: Freeing unused kernel memory: 39552K Apr 23 23:18:02.051105 kernel: Run /init as init process Apr 23 23:18:02.051109 kernel: with arguments: Apr 23 23:18:02.051114 kernel: /init Apr 23 23:18:02.051118 kernel: with environment: Apr 23 23:18:02.051123 kernel: HOME=/ Apr 23 23:18:02.051127 kernel: TERM=linux Apr 23 23:18:02.051133 systemd[1]: Successfully made /usr/ read-only. Apr 23 23:18:02.051141 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 23 23:18:02.051147 systemd[1]: Detected virtualization microsoft. Apr 23 23:18:02.051151 systemd[1]: Detected architecture arm64. Apr 23 23:18:02.051156 systemd[1]: Running in initrd. Apr 23 23:18:02.051161 systemd[1]: No hostname configured, using default hostname. Apr 23 23:18:02.051167 systemd[1]: Hostname set to . Apr 23 23:18:02.051172 systemd[1]: Initializing machine ID from random generator. Apr 23 23:18:02.051178 systemd[1]: Queued start job for default target initrd.target. Apr 23 23:18:02.051183 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 23 23:18:02.051188 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 23 23:18:02.051194 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 23 23:18:02.051199 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 23 23:18:02.051204 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 23 23:18:02.051210 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 23 23:18:02.051217 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 23 23:18:02.051222 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 23 23:18:02.051227 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 23 23:18:02.051233 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 23 23:18:02.051238 systemd[1]: Reached target paths.target - Path Units. Apr 23 23:18:02.051243 systemd[1]: Reached target slices.target - Slice Units. Apr 23 23:18:02.051248 systemd[1]: Reached target swap.target - Swaps. Apr 23 23:18:02.051253 systemd[1]: Reached target timers.target - Timer Units. Apr 23 23:18:02.051259 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 23 23:18:02.051264 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 23 23:18:02.051269 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 23 23:18:02.051275 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Apr 23 23:18:02.051280 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 23 23:18:02.051285 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 23 23:18:02.051290 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 23 23:18:02.051295 systemd[1]: Reached target sockets.target - Socket Units. Apr 23 23:18:02.051301 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 23 23:18:02.051306 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 23 23:18:02.051312 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 23 23:18:02.051317 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Apr 23 23:18:02.051322 systemd[1]: Starting systemd-fsck-usr.service... Apr 23 23:18:02.051327 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 23 23:18:02.051333 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 23 23:18:02.051348 systemd-journald[226]: Collecting audit messages is disabled. Apr 23 23:18:02.051362 systemd-journald[226]: Journal started Apr 23 23:18:02.051376 systemd-journald[226]: Runtime Journal (/run/log/journal/9b5bd2eaece94edcaa5acf52eb1cc645) is 8M, max 78.3M, 70.3M free. Apr 23 23:18:02.054847 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:18:02.059507 systemd-modules-load[228]: Inserted module 'overlay' Apr 23 23:18:02.085619 systemd[1]: Started systemd-journald.service - Journal Service. Apr 23 23:18:02.085651 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 23 23:18:02.087888 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 23 23:18:02.093677 kernel: Bridge firewalling registered Apr 23 23:18:02.089035 systemd-modules-load[228]: Inserted module 'br_netfilter' Apr 23 23:18:02.102677 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 23 23:18:02.109004 systemd[1]: Finished systemd-fsck-usr.service. Apr 23 23:18:02.120356 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 23 23:18:02.124804 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:02.135112 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 23 23:18:02.153903 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 23 23:18:02.162928 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 23 23:18:02.175149 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 23 23:18:02.186764 systemd-tmpfiles[251]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Apr 23 23:18:02.189754 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:18:02.204865 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 23 23:18:02.211111 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 23 23:18:02.222998 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 23 23:18:02.240065 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 23 23:18:02.248965 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 23 23:18:02.259797 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 23 23:18:02.274972 dracut-cmdline[263]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=8669c84e6bfac0c003f3ced682d9b5c0fda27fc2948639441be65941607b4c3d Apr 23 23:18:02.304957 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 23 23:18:02.312791 systemd-resolved[264]: Positive Trust Anchors: Apr 23 23:18:02.312800 systemd-resolved[264]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 23 23:18:02.312839 systemd-resolved[264]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 23 23:18:02.314400 systemd-resolved[264]: Defaulting to hostname 'linux'. Apr 23 23:18:02.315877 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 23 23:18:02.326327 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 23 23:18:02.390834 kernel: SCSI subsystem initialized Apr 23 23:18:02.396833 kernel: Loading iSCSI transport class v2.0-870. Apr 23 23:18:02.403840 kernel: iscsi: registered transport (tcp) Apr 23 23:18:02.417144 kernel: iscsi: registered transport (qla4xxx) Apr 23 23:18:02.417154 kernel: QLogic iSCSI HBA Driver Apr 23 23:18:02.430505 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 23 23:18:02.451045 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 23 23:18:02.457441 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 23 23:18:02.505734 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 23 23:18:02.514967 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 23 23:18:02.567833 kernel: raid6: neonx8 gen() 18556 MB/s Apr 23 23:18:02.586831 kernel: raid6: neonx4 gen() 18556 MB/s Apr 23 23:18:02.605828 kernel: raid6: neonx2 gen() 17100 MB/s Apr 23 23:18:02.625828 kernel: raid6: neonx1 gen() 15020 MB/s Apr 23 23:18:02.644843 kernel: raid6: int64x8 gen() 10530 MB/s Apr 23 23:18:02.663829 kernel: raid6: int64x4 gen() 10615 MB/s Apr 23 23:18:02.683829 kernel: raid6: int64x2 gen() 8980 MB/s Apr 23 23:18:02.705195 kernel: raid6: int64x1 gen() 7032 MB/s Apr 23 23:18:02.705237 kernel: raid6: using algorithm neonx8 gen() 18556 MB/s Apr 23 23:18:02.727115 kernel: raid6: .... xor() 14893 MB/s, rmw enabled Apr 23 23:18:02.727152 kernel: raid6: using neon recovery algorithm Apr 23 23:18:02.735917 kernel: xor: measuring software checksum speed Apr 23 23:18:02.735926 kernel: 8regs : 28675 MB/sec Apr 23 23:18:02.738432 kernel: 32regs : 28830 MB/sec Apr 23 23:18:02.740928 kernel: arm64_neon : 37597 MB/sec Apr 23 23:18:02.743867 kernel: xor: using function: arm64_neon (37597 MB/sec) Apr 23 23:18:02.782846 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 23 23:18:02.787493 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 23 23:18:02.796940 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 23 23:18:02.820629 systemd-udevd[475]: Using default interface naming scheme 'v255'. Apr 23 23:18:02.824630 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 23 23:18:02.836456 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 23 23:18:02.858720 dracut-pre-trigger[485]: rd.md=0: removing MD RAID activation Apr 23 23:18:02.877775 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 23 23:18:02.888658 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 23 23:18:02.925941 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 23 23:18:02.936863 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 23 23:18:02.994844 kernel: hv_vmbus: Vmbus version:5.3 Apr 23 23:18:03.014904 kernel: hv_vmbus: registering driver hid_hyperv Apr 23 23:18:03.014937 kernel: pps_core: LinuxPPS API ver. 1 registered Apr 23 23:18:03.014946 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Apr 23 23:18:03.010211 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:18:03.027143 kernel: PTP clock support registered Apr 23 23:18:03.010321 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:03.041705 kernel: hv_utils: Registering HyperV Utility Driver Apr 23 23:18:03.027060 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:18:03.037646 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:18:02.652647 kernel: hv_vmbus: registering driver hv_utils Apr 23 23:18:02.680562 kernel: hv_vmbus: registering driver hyperv_keyboard Apr 23 23:18:02.680576 kernel: hv_utils: Shutdown IC version 3.2 Apr 23 23:18:02.680582 kernel: hv_utils: TimeSync IC version 4.0 Apr 23 23:18:02.680588 kernel: hv_utils: Heartbeat IC version 3.0 Apr 23 23:18:02.686259 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input0 Apr 23 23:18:02.686275 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input1 Apr 23 23:18:02.686283 systemd-journald[226]: Time jumped backwards, rotating. Apr 23 23:18:02.686324 kernel: hv_vmbus: registering driver hv_netvsc Apr 23 23:18:02.686330 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Apr 23 23:18:02.652407 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Apr 23 23:18:02.660292 systemd-resolved[264]: Clock change detected. Flushing caches. Apr 23 23:18:02.664612 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:18:02.664678 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:02.710815 kernel: hv_vmbus: registering driver hv_storvsc Apr 23 23:18:02.710827 kernel: scsi host1: storvsc_host_t Apr 23 23:18:02.681226 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:18:02.717942 kernel: scsi host0: storvsc_host_t Apr 23 23:18:02.720427 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Apr 23 23:18:02.729051 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 5 Apr 23 23:18:02.739622 kernel: hv_netvsc 000d3af5-9260-000d-3af5-9260000d3af5 eth0: VF slot 1 added Apr 23 23:18:02.741910 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:02.767958 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Apr 23 23:18:02.768091 kernel: hv_vmbus: registering driver hv_pci Apr 23 23:18:02.768099 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Apr 23 23:18:02.768163 kernel: hv_pci e5fab05c-768b-4187-bbe5-b56a3a979058: PCI VMBus probing: Using version 0x10004 Apr 23 23:18:02.768228 kernel: sd 0:0:0:0: [sda] Write Protect is off Apr 23 23:18:02.769614 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Apr 23 23:18:02.779004 kernel: hv_pci e5fab05c-768b-4187-bbe5-b56a3a979058: PCI host bridge to bus 768b:00 Apr 23 23:18:02.779120 kernel: pci_bus 768b:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Apr 23 23:18:02.779222 kernel: pci_bus 768b:00: No busn resource found for root bus, will use [bus 00-ff] Apr 23 23:18:02.788625 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Apr 23 23:18:02.794627 kernel: pci 768b:00:02.0: [15b3:101a] type 00 class 0x020000 PCIe Endpoint Apr 23 23:18:02.799615 kernel: pci 768b:00:02.0: BAR 0 [mem 0xfc0000000-0xfc00fffff 64bit pref] Apr 23 23:18:02.804621 kernel: pci 768b:00:02.0: enabling Extended Tags Apr 23 23:18:02.818629 kernel: pci 768b:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 768b:00:02.0 (capable of 252.048 Gb/s with 16.0 GT/s PCIe x16 link) Apr 23 23:18:02.827775 kernel: pci_bus 768b:00: busn_res: [bus 00-ff] end is updated to 00 Apr 23 23:18:02.827886 kernel: pci 768b:00:02.0: BAR 0 [mem 0xfc0000000-0xfc00fffff 64bit pref]: assigned Apr 23 23:18:02.846934 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 23 23:18:02.846956 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Apr 23 23:18:02.854819 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Apr 23 23:18:02.854953 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 23 23:18:02.856618 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Apr 23 23:18:02.871603 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#99 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Apr 23 23:18:02.894639 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#67 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Apr 23 23:18:02.911691 kernel: mlx5_core 768b:00:02.0: enabling device (0000 -> 0002) Apr 23 23:18:02.919180 kernel: mlx5_core 768b:00:02.0: PTM is not supported by PCIe Apr 23 23:18:02.919306 kernel: mlx5_core 768b:00:02.0: firmware version: 16.30.5026 Apr 23 23:18:03.086226 kernel: hv_netvsc 000d3af5-9260-000d-3af5-9260000d3af5 eth0: VF registering: eth1 Apr 23 23:18:03.086390 kernel: mlx5_core 768b:00:02.0 eth1: joined to eth0 Apr 23 23:18:03.092615 kernel: mlx5_core 768b:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Apr 23 23:18:03.100988 kernel: mlx5_core 768b:00:02.0 enP30347s1: renamed from eth1 Apr 23 23:18:03.315368 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Apr 23 23:18:03.358741 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Apr 23 23:18:03.423376 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Apr 23 23:18:03.448204 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Apr 23 23:18:03.453669 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Apr 23 23:18:03.465615 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 23 23:18:03.474634 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 23 23:18:03.483453 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 23 23:18:03.493140 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 23 23:18:03.502146 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 23 23:18:03.517416 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 23 23:18:03.532294 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 23 23:18:03.547671 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 23 23:18:04.566615 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 23 23:18:04.567612 disk-uuid[663]: The operation has completed successfully. Apr 23 23:18:04.637020 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 23 23:18:04.640488 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 23 23:18:04.666627 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 23 23:18:04.686627 sh[821]: Success Apr 23 23:18:04.719510 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 23 23:18:04.719548 kernel: device-mapper: uevent: version 1.0.3 Apr 23 23:18:04.724512 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Apr 23 23:18:04.734609 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Apr 23 23:18:04.995252 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 23 23:18:05.005902 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 23 23:18:05.010941 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 23 23:18:05.037051 kernel: BTRFS: device fsid 2db32ba8-c7e9-4b6a-ba75-58982c25581e devid 1 transid 32 /dev/mapper/usr (254:0) scanned by mount (839) Apr 23 23:18:05.037072 kernel: BTRFS info (device dm-0): first mount of filesystem 2db32ba8-c7e9-4b6a-ba75-58982c25581e Apr 23 23:18:05.041488 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Apr 23 23:18:05.342633 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Apr 23 23:18:05.342694 kernel: BTRFS info (device dm-0 state E): enabling free space tree Apr 23 23:18:05.386222 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 23 23:18:05.389792 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Apr 23 23:18:05.397229 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 23 23:18:05.397809 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 23 23:18:05.417228 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 23 23:18:05.448608 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (873) Apr 23 23:18:05.458428 kernel: BTRFS info (device sda6): first mount of filesystem a3954155-494f-4049-93fc-7ec9255747d0 Apr 23 23:18:05.458455 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 23 23:18:05.484545 kernel: BTRFS info (device sda6): turning on async discard Apr 23 23:18:05.484576 kernel: BTRFS info (device sda6): enabling free space tree Apr 23 23:18:05.492639 kernel: BTRFS info (device sda6): last unmount of filesystem a3954155-494f-4049-93fc-7ec9255747d0 Apr 23 23:18:05.493408 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 23 23:18:05.499275 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 23 23:18:05.528618 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 23 23:18:05.539789 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 23 23:18:05.569880 systemd-networkd[1008]: lo: Link UP Apr 23 23:18:05.569891 systemd-networkd[1008]: lo: Gained carrier Apr 23 23:18:05.570562 systemd-networkd[1008]: Enumeration completed Apr 23 23:18:05.572647 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 23 23:18:05.577086 systemd-networkd[1008]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:18:05.577088 systemd-networkd[1008]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 23 23:18:05.579697 systemd[1]: Reached target network.target - Network. Apr 23 23:18:05.647611 kernel: mlx5_core 768b:00:02.0 enP30347s1: Link up Apr 23 23:18:05.678940 kernel: hv_netvsc 000d3af5-9260-000d-3af5-9260000d3af5 eth0: Data path switched to VF: enP30347s1 Apr 23 23:18:05.678629 systemd-networkd[1008]: enP30347s1: Link UP Apr 23 23:18:05.678685 systemd-networkd[1008]: eth0: Link UP Apr 23 23:18:05.678802 systemd-networkd[1008]: eth0: Gained carrier Apr 23 23:18:05.678825 systemd-networkd[1008]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:18:05.696102 systemd-networkd[1008]: enP30347s1: Gained carrier Apr 23 23:18:05.707620 systemd-networkd[1008]: eth0: DHCPv4 address 10.0.0.13/24, gateway 10.0.0.1 acquired from 168.63.129.16 Apr 23 23:18:06.477751 ignition[979]: Ignition 2.22.0 Apr 23 23:18:06.477762 ignition[979]: Stage: fetch-offline Apr 23 23:18:06.480750 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 23 23:18:06.477870 ignition[979]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:18:06.489299 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 23 23:18:06.477877 ignition[979]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 23 23:18:06.477963 ignition[979]: parsed url from cmdline: "" Apr 23 23:18:06.477966 ignition[979]: no config URL provided Apr 23 23:18:06.477971 ignition[979]: reading system config file "/usr/lib/ignition/user.ign" Apr 23 23:18:06.477976 ignition[979]: no config at "/usr/lib/ignition/user.ign" Apr 23 23:18:06.477980 ignition[979]: failed to fetch config: resource requires networking Apr 23 23:18:06.478092 ignition[979]: Ignition finished successfully Apr 23 23:18:06.531108 ignition[1019]: Ignition 2.22.0 Apr 23 23:18:06.531112 ignition[1019]: Stage: fetch Apr 23 23:18:06.531299 ignition[1019]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:18:06.531307 ignition[1019]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 23 23:18:06.531364 ignition[1019]: parsed url from cmdline: "" Apr 23 23:18:06.531367 ignition[1019]: no config URL provided Apr 23 23:18:06.531370 ignition[1019]: reading system config file "/usr/lib/ignition/user.ign" Apr 23 23:18:06.531375 ignition[1019]: no config at "/usr/lib/ignition/user.ign" Apr 23 23:18:06.531390 ignition[1019]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Apr 23 23:18:06.644250 ignition[1019]: GET result: OK Apr 23 23:18:06.644301 ignition[1019]: config has been read from IMDS userdata Apr 23 23:18:06.644322 ignition[1019]: parsing config with SHA512: 17c2c217f55bb6e3ce10b5f8c59958ceb844705bf095be9415b041bc62bba12ec749429a4a3cbe1a18dc7341c1302d7e6b48b14556eaf5682cbfb923467aff6d Apr 23 23:18:06.650445 unknown[1019]: fetched base config from "system" Apr 23 23:18:06.650456 unknown[1019]: fetched base config from "system" Apr 23 23:18:06.651686 ignition[1019]: fetch: fetch complete Apr 23 23:18:06.650460 unknown[1019]: fetched user config from "azure" Apr 23 23:18:06.651691 ignition[1019]: fetch: fetch passed Apr 23 23:18:06.653372 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 23 23:18:06.651732 ignition[1019]: Ignition finished successfully Apr 23 23:18:06.658566 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 23 23:18:06.695734 ignition[1026]: Ignition 2.22.0 Apr 23 23:18:06.695745 ignition[1026]: Stage: kargs Apr 23 23:18:06.698253 ignition[1026]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:18:06.703631 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 23 23:18:06.698262 ignition[1026]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 23 23:18:06.711440 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 23 23:18:06.698790 ignition[1026]: kargs: kargs passed Apr 23 23:18:06.698824 ignition[1026]: Ignition finished successfully Apr 23 23:18:06.737073 ignition[1032]: Ignition 2.22.0 Apr 23 23:18:06.737093 ignition[1032]: Stage: disks Apr 23 23:18:06.742555 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 23 23:18:06.737257 ignition[1032]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:18:06.746688 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 23 23:18:06.737263 ignition[1032]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 23 23:18:06.754586 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 23 23:18:06.737796 ignition[1032]: disks: disks passed Apr 23 23:18:06.762646 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 23 23:18:06.737832 ignition[1032]: Ignition finished successfully Apr 23 23:18:06.770922 systemd[1]: Reached target sysinit.target - System Initialization. Apr 23 23:18:06.779305 systemd[1]: Reached target basic.target - Basic System. Apr 23 23:18:06.788548 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 23 23:18:06.868164 systemd-fsck[1040]: ROOT: clean, 15/7326000 files, 477845/7359488 blocks Apr 23 23:18:06.875271 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 23 23:18:06.881137 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 23 23:18:07.115635 kernel: EXT4-fs (sda9): mounted filesystem 753efcb9-de86-4e47-981f-2dbd4690452d r/w with ordered data mode. Quota mode: none. Apr 23 23:18:07.115841 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 23 23:18:07.122451 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 23 23:18:07.145649 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 23 23:18:07.163672 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 23 23:18:07.171082 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 23 23:18:07.176326 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 23 23:18:07.176350 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 23 23:18:07.187314 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 23 23:18:07.207534 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 23 23:18:07.224610 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (1054) Apr 23 23:18:07.234931 kernel: BTRFS info (device sda6): first mount of filesystem a3954155-494f-4049-93fc-7ec9255747d0 Apr 23 23:18:07.234958 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 23 23:18:07.243956 kernel: BTRFS info (device sda6): turning on async discard Apr 23 23:18:07.243985 kernel: BTRFS info (device sda6): enabling free space tree Apr 23 23:18:07.244913 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 23 23:18:07.276794 systemd-networkd[1008]: eth0: Gained IPv6LL Apr 23 23:18:07.706043 coreos-metadata[1056]: Apr 23 23:18:07.706 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Apr 23 23:18:07.712020 coreos-metadata[1056]: Apr 23 23:18:07.711 INFO Fetch successful Apr 23 23:18:07.712020 coreos-metadata[1056]: Apr 23 23:18:07.711 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Apr 23 23:18:07.724308 coreos-metadata[1056]: Apr 23 23:18:07.724 INFO Fetch successful Apr 23 23:18:07.739209 coreos-metadata[1056]: Apr 23 23:18:07.739 INFO wrote hostname ci-4459.2.4-n-8b75ae451f to /sysroot/etc/hostname Apr 23 23:18:07.745974 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 23 23:18:07.969983 initrd-setup-root[1085]: cut: /sysroot/etc/passwd: No such file or directory Apr 23 23:18:07.991111 initrd-setup-root[1092]: cut: /sysroot/etc/group: No such file or directory Apr 23 23:18:07.997637 initrd-setup-root[1099]: cut: /sysroot/etc/shadow: No such file or directory Apr 23 23:18:08.004230 initrd-setup-root[1106]: cut: /sysroot/etc/gshadow: No such file or directory Apr 23 23:18:08.868164 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 23 23:18:08.873394 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 23 23:18:08.890165 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 23 23:18:08.899940 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 23 23:18:08.909553 kernel: BTRFS info (device sda6): last unmount of filesystem a3954155-494f-4049-93fc-7ec9255747d0 Apr 23 23:18:08.931618 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 23 23:18:08.941617 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 23 23:18:08.954236 ignition[1174]: INFO : Ignition 2.22.0 Apr 23 23:18:08.954236 ignition[1174]: INFO : Stage: mount Apr 23 23:18:08.954236 ignition[1174]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 23 23:18:08.954236 ignition[1174]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 23 23:18:08.954236 ignition[1174]: INFO : mount: mount passed Apr 23 23:18:08.954236 ignition[1174]: INFO : Ignition finished successfully Apr 23 23:18:08.946806 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 23 23:18:08.966686 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 23 23:18:08.997613 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (1186) Apr 23 23:18:08.997638 kernel: BTRFS info (device sda6): first mount of filesystem a3954155-494f-4049-93fc-7ec9255747d0 Apr 23 23:18:09.006926 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 23 23:18:09.015504 kernel: BTRFS info (device sda6): turning on async discard Apr 23 23:18:09.015523 kernel: BTRFS info (device sda6): enabling free space tree Apr 23 23:18:09.017230 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 23 23:18:09.043944 ignition[1204]: INFO : Ignition 2.22.0 Apr 23 23:18:09.043944 ignition[1204]: INFO : Stage: files Apr 23 23:18:09.049770 ignition[1204]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 23 23:18:09.049770 ignition[1204]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 23 23:18:09.049770 ignition[1204]: DEBUG : files: compiled without relabeling support, skipping Apr 23 23:18:09.073886 ignition[1204]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 23 23:18:09.073886 ignition[1204]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 23 23:18:09.131304 ignition[1204]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 23 23:18:09.136644 ignition[1204]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 23 23:18:09.136644 ignition[1204]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 23 23:18:09.131642 unknown[1204]: wrote ssh authorized keys file for user: core Apr 23 23:18:09.162735 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Apr 23 23:18:09.170477 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Apr 23 23:18:09.231861 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 23 23:18:09.506989 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Apr 23 23:18:09.506989 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Apr 23 23:18:09.521687 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.4-arm64.raw: attempt #1 Apr 23 23:18:09.911808 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 23 23:18:10.248604 ignition[1204]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Apr 23 23:18:10.248604 ignition[1204]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 23 23:18:10.281875 ignition[1204]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 23 23:18:10.290266 ignition[1204]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 23 23:18:10.290266 ignition[1204]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 23 23:18:10.290266 ignition[1204]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Apr 23 23:18:10.290266 ignition[1204]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Apr 23 23:18:10.290266 ignition[1204]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 23 23:18:10.290266 ignition[1204]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 23 23:18:10.290266 ignition[1204]: INFO : files: files passed Apr 23 23:18:10.290266 ignition[1204]: INFO : Ignition finished successfully Apr 23 23:18:10.292801 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 23 23:18:10.302533 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 23 23:18:10.326068 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 23 23:18:10.380915 initrd-setup-root-after-ignition[1235]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 23 23:18:10.345953 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 23 23:18:10.397743 initrd-setup-root-after-ignition[1231]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 23 23:18:10.397743 initrd-setup-root-after-ignition[1231]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 23 23:18:10.346017 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 23 23:18:10.357093 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 23 23:18:10.365823 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 23 23:18:10.374591 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 23 23:18:10.412159 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 23 23:18:10.412236 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 23 23:18:10.420711 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 23 23:18:10.429485 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 23 23:18:10.437514 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 23 23:18:10.438054 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 23 23:18:10.479008 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 23 23:18:10.484448 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 23 23:18:10.508467 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 23 23:18:10.513325 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 23 23:18:10.522730 systemd[1]: Stopped target timers.target - Timer Units. Apr 23 23:18:10.531653 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 23 23:18:10.531740 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 23 23:18:10.543488 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 23 23:18:10.547857 systemd[1]: Stopped target basic.target - Basic System. Apr 23 23:18:10.556078 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 23 23:18:10.564139 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 23 23:18:10.572164 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 23 23:18:10.580695 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Apr 23 23:18:10.589550 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 23 23:18:10.597888 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 23 23:18:10.607011 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 23 23:18:10.614878 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 23 23:18:10.623859 systemd[1]: Stopped target swap.target - Swaps. Apr 23 23:18:10.631047 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 23 23:18:10.631134 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 23 23:18:10.641853 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 23 23:18:10.646461 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 23 23:18:10.655004 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 23 23:18:10.658828 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 23 23:18:10.664168 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 23 23:18:10.664239 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 23 23:18:10.676666 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 23 23:18:10.676741 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 23 23:18:10.681947 systemd[1]: ignition-files.service: Deactivated successfully. Apr 23 23:18:10.682016 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 23 23:18:10.689707 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 23 23:18:10.689769 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 23 23:18:10.700620 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 23 23:18:10.773561 ignition[1256]: INFO : Ignition 2.22.0 Apr 23 23:18:10.773561 ignition[1256]: INFO : Stage: umount Apr 23 23:18:10.773561 ignition[1256]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 23 23:18:10.773561 ignition[1256]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 23 23:18:10.773561 ignition[1256]: INFO : umount: umount passed Apr 23 23:18:10.773561 ignition[1256]: INFO : Ignition finished successfully Apr 23 23:18:10.713468 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 23 23:18:10.713579 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 23 23:18:10.723565 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 23 23:18:10.741964 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 23 23:18:10.742076 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 23 23:18:10.751623 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 23 23:18:10.751703 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 23 23:18:10.775002 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 23 23:18:10.775553 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 23 23:18:10.775641 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 23 23:18:10.782706 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 23 23:18:10.782911 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 23 23:18:10.788991 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 23 23:18:10.789031 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 23 23:18:10.796788 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 23 23:18:10.796822 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 23 23:18:10.807383 systemd[1]: Stopped target network.target - Network. Apr 23 23:18:10.814226 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 23 23:18:10.814276 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 23 23:18:10.823055 systemd[1]: Stopped target paths.target - Path Units. Apr 23 23:18:10.831357 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 23 23:18:10.835890 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 23 23:18:10.840781 systemd[1]: Stopped target slices.target - Slice Units. Apr 23 23:18:10.847829 systemd[1]: Stopped target sockets.target - Socket Units. Apr 23 23:18:10.855445 systemd[1]: iscsid.socket: Deactivated successfully. Apr 23 23:18:10.855477 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 23 23:18:10.863437 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 23 23:18:10.863464 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 23 23:18:10.871684 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 23 23:18:10.871722 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 23 23:18:10.879583 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 23 23:18:10.879614 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 23 23:18:10.888327 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 23 23:18:10.892744 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 23 23:18:10.912784 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 23 23:18:10.912859 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 23 23:18:10.922130 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Apr 23 23:18:10.922302 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 23 23:18:10.922401 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 23 23:18:11.117020 kernel: hv_netvsc 000d3af5-9260-000d-3af5-9260000d3af5 eth0: Data path switched from VF: enP30347s1 Apr 23 23:18:10.933411 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Apr 23 23:18:10.933574 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 23 23:18:10.936516 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 23 23:18:10.942232 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 23 23:18:10.943616 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 23 23:18:10.951333 systemd[1]: Stopped target network-pre.target - Preparation for Network. Apr 23 23:18:10.957829 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 23 23:18:10.957875 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 23 23:18:10.966975 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 23 23:18:10.967165 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 23 23:18:10.979717 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 23 23:18:10.994413 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 23 23:18:10.994469 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 23 23:18:10.999363 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 23 23:18:10.999394 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:18:11.009929 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 23 23:18:11.009958 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 23 23:18:11.014294 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 23 23:18:11.014320 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 23 23:18:11.025931 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 23 23:18:11.033280 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Apr 23 23:18:11.033324 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Apr 23 23:18:11.054448 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 23 23:18:11.054557 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 23 23:18:11.064522 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 23 23:18:11.064553 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 23 23:18:11.073060 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 23 23:18:11.073089 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 23 23:18:11.080958 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 23 23:18:11.080995 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 23 23:18:11.089201 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 23 23:18:11.089234 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 23 23:18:11.104515 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 23 23:18:11.104556 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 23 23:18:11.122726 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 23 23:18:11.136921 systemd[1]: systemd-network-generator.service: Deactivated successfully. Apr 23 23:18:11.136973 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Apr 23 23:18:11.148645 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 23 23:18:11.148687 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 23 23:18:11.157974 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:18:11.158009 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:11.166986 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Apr 23 23:18:11.167026 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Apr 23 23:18:11.167052 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Apr 23 23:18:11.167248 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 23 23:18:11.167316 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 23 23:18:11.212477 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 23 23:18:11.212581 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 23 23:18:11.367736 systemd-journald[226]: Received SIGTERM from PID 1 (systemd). Apr 23 23:18:11.220908 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 23 23:18:11.231365 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 23 23:18:11.261571 systemd[1]: Switching root. Apr 23 23:18:11.378084 systemd-journald[226]: Journal stopped Apr 23 23:18:15.890587 kernel: SELinux: policy capability network_peer_controls=1 Apr 23 23:18:15.890614 kernel: SELinux: policy capability open_perms=1 Apr 23 23:18:15.890621 kernel: SELinux: policy capability extended_socket_class=1 Apr 23 23:18:15.890626 kernel: SELinux: policy capability always_check_network=0 Apr 23 23:18:15.890631 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 23 23:18:15.890638 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 23 23:18:15.890644 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 23 23:18:15.890649 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 23 23:18:15.890654 kernel: SELinux: policy capability userspace_initial_context=0 Apr 23 23:18:15.890660 kernel: audit: type=1403 audit(1776986292.307:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 23 23:18:15.890666 systemd[1]: Successfully loaded SELinux policy in 171.148ms. Apr 23 23:18:15.890674 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 4.342ms. Apr 23 23:18:15.890681 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 23 23:18:15.890686 systemd[1]: Detected virtualization microsoft. Apr 23 23:18:15.890693 systemd[1]: Detected architecture arm64. Apr 23 23:18:15.890699 systemd[1]: Detected first boot. Apr 23 23:18:15.890706 systemd[1]: Hostname set to . Apr 23 23:18:15.890711 systemd[1]: Initializing machine ID from random generator. Apr 23 23:18:15.890717 zram_generator::config[1298]: No configuration found. Apr 23 23:18:15.890724 kernel: NET: Registered PF_VSOCK protocol family Apr 23 23:18:15.890730 systemd[1]: Populated /etc with preset unit settings. Apr 23 23:18:15.890736 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Apr 23 23:18:15.890742 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 23 23:18:15.890749 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 23 23:18:15.890755 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 23 23:18:15.890761 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 23 23:18:15.890768 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 23 23:18:15.890774 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 23 23:18:15.890780 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 23 23:18:15.890786 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 23 23:18:15.890793 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 23 23:18:15.890799 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 23 23:18:15.890805 systemd[1]: Created slice user.slice - User and Session Slice. Apr 23 23:18:15.890811 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 23 23:18:15.890818 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 23 23:18:15.890824 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 23 23:18:15.890830 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 23 23:18:15.890836 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 23 23:18:15.890843 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 23 23:18:15.890849 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Apr 23 23:18:15.890857 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 23 23:18:15.890863 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 23 23:18:15.890870 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 23 23:18:15.890876 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 23 23:18:15.890882 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 23 23:18:15.890888 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 23 23:18:15.890895 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 23 23:18:15.890901 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 23 23:18:15.890907 systemd[1]: Reached target slices.target - Slice Units. Apr 23 23:18:15.890913 systemd[1]: Reached target swap.target - Swaps. Apr 23 23:18:15.890919 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 23 23:18:15.890925 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 23 23:18:15.890933 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Apr 23 23:18:15.890939 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 23 23:18:15.890946 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 23 23:18:15.890952 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 23 23:18:15.890958 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 23 23:18:15.890964 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 23 23:18:15.890970 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 23 23:18:15.890977 systemd[1]: Mounting media.mount - External Media Directory... Apr 23 23:18:15.890983 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 23 23:18:15.890990 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 23 23:18:15.890996 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 23 23:18:15.891003 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 23 23:18:15.891009 systemd[1]: Reached target machines.target - Containers. Apr 23 23:18:15.891015 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 23 23:18:15.891021 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:18:15.891028 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 23 23:18:15.891034 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 23 23:18:15.891041 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:18:15.891047 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 23 23:18:15.891053 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 23 23:18:15.891059 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 23 23:18:15.891065 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 23 23:18:15.891072 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 23 23:18:15.891078 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 23 23:18:15.891085 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 23 23:18:15.891091 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 23 23:18:15.891097 systemd[1]: Stopped systemd-fsck-usr.service. Apr 23 23:18:15.891103 kernel: fuse: init (API version 7.41) Apr 23 23:18:15.891109 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:18:15.891116 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 23 23:18:15.891122 kernel: loop: module loaded Apr 23 23:18:15.891128 kernel: ACPI: bus type drm_connector registered Apr 23 23:18:15.891134 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 23 23:18:15.891141 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 23 23:18:15.891147 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 23 23:18:15.891164 systemd-journald[1378]: Collecting audit messages is disabled. Apr 23 23:18:15.891178 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Apr 23 23:18:15.891185 systemd-journald[1378]: Journal started Apr 23 23:18:15.891199 systemd-journald[1378]: Runtime Journal (/run/log/journal/bd1ffd94313c4b87a28badb6cd5996a1) is 8M, max 78.3M, 70.3M free. Apr 23 23:18:15.197234 systemd[1]: Queued start job for default target multi-user.target. Apr 23 23:18:15.205025 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Apr 23 23:18:15.205360 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 23 23:18:15.205617 systemd[1]: systemd-journald.service: Consumed 2.312s CPU time. Apr 23 23:18:15.916172 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 23 23:18:15.926261 systemd[1]: verity-setup.service: Deactivated successfully. Apr 23 23:18:15.926503 systemd[1]: Stopped verity-setup.service. Apr 23 23:18:15.939200 systemd[1]: Started systemd-journald.service - Journal Service. Apr 23 23:18:15.939762 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 23 23:18:15.944196 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 23 23:18:15.948607 systemd[1]: Mounted media.mount - External Media Directory. Apr 23 23:18:15.952427 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 23 23:18:15.956768 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 23 23:18:15.961368 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 23 23:18:15.965415 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 23 23:18:15.970269 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 23 23:18:15.975678 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 23 23:18:15.975796 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 23 23:18:15.980664 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:18:15.980783 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:18:15.985440 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 23 23:18:15.986672 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 23 23:18:15.992829 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 23 23:18:15.992946 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 23 23:18:15.997962 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 23 23:18:15.998069 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 23 23:18:16.002644 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 23 23:18:16.002761 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 23 23:18:16.007449 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 23 23:18:16.012131 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 23 23:18:16.017402 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 23 23:18:16.022712 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Apr 23 23:18:16.027961 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 23 23:18:16.040902 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 23 23:18:16.046253 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 23 23:18:16.055133 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 23 23:18:16.062164 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 23 23:18:16.062195 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 23 23:18:16.067074 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Apr 23 23:18:16.072897 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 23 23:18:16.076898 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:18:16.077635 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 23 23:18:16.082483 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 23 23:18:16.087097 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 23 23:18:16.087754 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 23 23:18:16.092091 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 23 23:18:16.094698 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 23 23:18:16.099290 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 23 23:18:16.105571 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 23 23:18:16.110750 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 23 23:18:16.115434 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 23 23:18:16.123256 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 23 23:18:16.129220 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 23 23:18:16.134712 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Apr 23 23:18:16.159122 systemd-journald[1378]: Time spent on flushing to /var/log/journal/bd1ffd94313c4b87a28badb6cd5996a1 is 69.597ms for 934 entries. Apr 23 23:18:16.159122 systemd-journald[1378]: System Journal (/var/log/journal/bd1ffd94313c4b87a28badb6cd5996a1) is 11.8M, max 2.6G, 2.6G free. Apr 23 23:18:16.318711 systemd-journald[1378]: Received client request to flush runtime journal. Apr 23 23:18:16.318741 kernel: loop0: detected capacity change from 0 to 119840 Apr 23 23:18:16.318749 systemd-journald[1378]: /var/log/journal/bd1ffd94313c4b87a28badb6cd5996a1/system.journal: Realtime clock jumped backwards relative to last journal entry, rotating. Apr 23 23:18:16.318762 systemd-journald[1378]: Rotating system journal. Apr 23 23:18:16.210668 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:18:16.285327 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 23 23:18:16.285835 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Apr 23 23:18:16.299287 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 23 23:18:16.304680 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 23 23:18:16.320648 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 23 23:18:16.435929 systemd-tmpfiles[1451]: ACLs are not supported, ignoring. Apr 23 23:18:16.435941 systemd-tmpfiles[1451]: ACLs are not supported, ignoring. Apr 23 23:18:16.440622 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 23 23:18:16.570623 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 23 23:18:16.630621 kernel: loop1: detected capacity change from 0 to 100632 Apr 23 23:18:16.689816 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 23 23:18:16.696578 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 23 23:18:16.722000 systemd-udevd[1459]: Using default interface naming scheme 'v255'. Apr 23 23:18:16.969186 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 23 23:18:16.979352 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 23 23:18:17.033164 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 23 23:18:17.063519 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Apr 23 23:18:17.082613 kernel: loop2: detected capacity change from 0 to 27936 Apr 23 23:18:17.096507 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 23 23:18:17.136611 kernel: mousedev: PS/2 mouse device common for all mice Apr 23 23:18:17.169109 kernel: hv_vmbus: registering driver hv_balloon Apr 23 23:18:17.169145 kernel: hv_vmbus: registering driver hyperv_fb Apr 23 23:18:17.169154 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Apr 23 23:18:17.174059 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Apr 23 23:18:17.178131 kernel: Console: switching to colour dummy device 80x25 Apr 23 23:18:17.184426 kernel: Console: switching to colour frame buffer device 128x48 Apr 23 23:18:17.201676 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Apr 23 23:18:17.201712 kernel: hv_balloon: Memory hot add disabled on ARM64 Apr 23 23:18:17.206717 systemd-networkd[1474]: lo: Link UP Apr 23 23:18:17.206725 systemd-networkd[1474]: lo: Gained carrier Apr 23 23:18:17.207916 systemd-networkd[1474]: Enumeration completed Apr 23 23:18:17.208145 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 23 23:18:17.212819 systemd-networkd[1474]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:18:17.212829 systemd-networkd[1474]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 23 23:18:17.217761 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Apr 23 23:18:17.228763 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 23 23:18:17.241683 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#256 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Apr 23 23:18:17.252838 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:18:17.262347 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:18:17.262498 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:17.270521 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:18:17.282680 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:18:17.283135 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:17.283617 kernel: mlx5_core 768b:00:02.0 enP30347s1: Link up Apr 23 23:18:17.291767 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:18:17.304612 kernel: hv_netvsc 000d3af5-9260-000d-3af5-9260000d3af5 eth0: Data path switched to VF: enP30347s1 Apr 23 23:18:17.307432 systemd-networkd[1474]: enP30347s1: Link UP Apr 23 23:18:17.307838 systemd-networkd[1474]: eth0: Link UP Apr 23 23:18:17.307879 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Apr 23 23:18:17.308161 systemd-networkd[1474]: eth0: Gained carrier Apr 23 23:18:17.309008 systemd-networkd[1474]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:18:17.316989 systemd-networkd[1474]: enP30347s1: Gained carrier Apr 23 23:18:17.323636 systemd-networkd[1474]: eth0: DHCPv4 address 10.0.0.13/24, gateway 10.0.0.1 acquired from 168.63.129.16 Apr 23 23:18:17.343630 kernel: MACsec IEEE 802.1AE Apr 23 23:18:17.408362 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Apr 23 23:18:17.417721 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 23 23:18:17.454626 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 23 23:18:17.457631 kernel: loop3: detected capacity change from 0 to 200864 Apr 23 23:18:17.497625 kernel: loop4: detected capacity change from 0 to 119840 Apr 23 23:18:17.509615 kernel: loop5: detected capacity change from 0 to 100632 Apr 23 23:18:17.521612 kernel: loop6: detected capacity change from 0 to 27936 Apr 23 23:18:17.533619 kernel: loop7: detected capacity change from 0 to 200864 Apr 23 23:18:17.545820 (sd-merge)[1606]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Apr 23 23:18:17.546183 (sd-merge)[1606]: Merged extensions into '/usr'. Apr 23 23:18:17.549941 systemd[1]: Reload requested from client PID 1437 ('systemd-sysext') (unit systemd-sysext.service)... Apr 23 23:18:17.549954 systemd[1]: Reloading... Apr 23 23:18:17.589641 zram_generator::config[1634]: No configuration found. Apr 23 23:18:17.761301 systemd[1]: Reloading finished in 211 ms. Apr 23 23:18:17.776511 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:18:17.781532 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 23 23:18:17.792423 systemd[1]: Starting ensure-sysext.service... Apr 23 23:18:17.798722 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 23 23:18:17.816660 systemd[1]: Reload requested from client PID 1694 ('systemctl') (unit ensure-sysext.service)... Apr 23 23:18:17.816672 systemd[1]: Reloading... Apr 23 23:18:17.818660 systemd-tmpfiles[1695]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Apr 23 23:18:17.818897 systemd-tmpfiles[1695]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Apr 23 23:18:17.819138 systemd-tmpfiles[1695]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 23 23:18:17.819373 systemd-tmpfiles[1695]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 23 23:18:17.819901 systemd-tmpfiles[1695]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 23 23:18:17.820519 systemd-tmpfiles[1695]: ACLs are not supported, ignoring. Apr 23 23:18:17.820658 systemd-tmpfiles[1695]: ACLs are not supported, ignoring. Apr 23 23:18:17.838085 systemd-tmpfiles[1695]: Detected autofs mount point /boot during canonicalization of boot. Apr 23 23:18:17.838157 systemd-tmpfiles[1695]: Skipping /boot Apr 23 23:18:17.845310 systemd-tmpfiles[1695]: Detected autofs mount point /boot during canonicalization of boot. Apr 23 23:18:17.846690 systemd-tmpfiles[1695]: Skipping /boot Apr 23 23:18:17.859618 zram_generator::config[1723]: No configuration found. Apr 23 23:18:18.029290 systemd[1]: Reloading finished in 212 ms. Apr 23 23:18:18.035531 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 23 23:18:18.056227 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 23 23:18:18.076282 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 23 23:18:18.081019 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:18:18.088796 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:18:18.094550 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 23 23:18:18.100831 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 23 23:18:18.106294 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:18:18.106538 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:18:18.110747 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 23 23:18:18.120798 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 23 23:18:18.128054 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 23 23:18:18.134077 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:18:18.138713 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:18:18.144373 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 23 23:18:18.144497 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 23 23:18:18.149471 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 23 23:18:18.149588 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 23 23:18:18.158314 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:18:18.162233 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:18:18.169827 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 23 23:18:18.181836 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 23 23:18:18.186360 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:18:18.186450 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:18:18.190313 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 23 23:18:18.198244 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 23 23:18:18.203607 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:18:18.203724 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:18:18.208627 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 23 23:18:18.208732 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 23 23:18:18.214061 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 23 23:18:18.214172 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 23 23:18:18.225832 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:18:18.228762 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:18:18.233890 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 23 23:18:18.244309 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 23 23:18:18.251753 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 23 23:18:18.256148 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:18:18.256238 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:18:18.256337 systemd[1]: Reached target time-set.target - System Time Set. Apr 23 23:18:18.261195 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:18:18.261327 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:18:18.266018 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 23 23:18:18.266132 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 23 23:18:18.270910 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 23 23:18:18.271025 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 23 23:18:18.273169 systemd-resolved[1790]: Positive Trust Anchors: Apr 23 23:18:18.273370 systemd-resolved[1790]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 23 23:18:18.273426 systemd-resolved[1790]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 23 23:18:18.277276 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 23 23:18:18.277384 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 23 23:18:18.284258 systemd[1]: Finished ensure-sysext.service. Apr 23 23:18:18.289453 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 23 23:18:18.289505 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 23 23:18:18.290201 systemd-resolved[1790]: Using system hostname 'ci-4459.2.4-n-8b75ae451f'. Apr 23 23:18:18.291210 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 23 23:18:18.295841 systemd[1]: Reached target network.target - Network. Apr 23 23:18:18.299626 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 23 23:18:18.323850 augenrules[1831]: No rules Apr 23 23:18:18.324664 systemd[1]: audit-rules.service: Deactivated successfully. Apr 23 23:18:18.324952 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 23 23:18:18.668760 systemd-networkd[1474]: eth0: Gained IPv6LL Apr 23 23:18:18.670858 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 23 23:18:18.676887 systemd[1]: Reached target network-online.target - Network is Online. Apr 23 23:18:18.915209 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 23 23:18:18.920778 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 23 23:18:21.666408 ldconfig[1432]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 23 23:18:21.679242 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 23 23:18:21.687741 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 23 23:18:21.704749 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 23 23:18:21.710756 systemd[1]: Reached target sysinit.target - System Initialization. Apr 23 23:18:21.715365 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 23 23:18:21.720975 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 23 23:18:21.726338 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 23 23:18:21.730891 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 23 23:18:21.736928 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 23 23:18:21.742527 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 23 23:18:21.742550 systemd[1]: Reached target paths.target - Path Units. Apr 23 23:18:21.746455 systemd[1]: Reached target timers.target - Timer Units. Apr 23 23:18:21.751944 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 23 23:18:21.757661 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 23 23:18:21.763022 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Apr 23 23:18:21.768931 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Apr 23 23:18:21.774580 systemd[1]: Reached target ssh-access.target - SSH Access Available. Apr 23 23:18:21.781913 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 23 23:18:21.786487 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Apr 23 23:18:21.792091 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 23 23:18:21.797058 systemd[1]: Reached target sockets.target - Socket Units. Apr 23 23:18:21.801654 systemd[1]: Reached target basic.target - Basic System. Apr 23 23:18:21.805773 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 23 23:18:21.805792 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 23 23:18:21.826961 systemd[1]: Starting chronyd.service - NTP client/server... Apr 23 23:18:21.839547 systemd[1]: Starting containerd.service - containerd container runtime... Apr 23 23:18:21.846771 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 23 23:18:21.853018 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 23 23:18:21.860801 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 23 23:18:21.872707 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 23 23:18:21.878769 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 23 23:18:21.884912 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 23 23:18:21.886711 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Apr 23 23:18:21.892165 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Apr 23 23:18:21.892996 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:18:21.900720 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 23 23:18:21.906776 jq[1852]: false Apr 23 23:18:21.909724 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 23 23:18:21.916023 KVP[1854]: KVP starting; pid is:1854 Apr 23 23:18:21.918164 chronyd[1844]: chronyd version 4.7 starting (+CMDMON +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +NTS +SECHASH +IPV6 -DEBUG) Apr 23 23:18:21.918542 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 23 23:18:21.920438 KVP[1854]: KVP LIC Version: 3.1 Apr 23 23:18:21.920639 kernel: hv_utils: KVP IC version 4.0 Apr 23 23:18:21.926747 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 23 23:18:21.935568 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 23 23:18:21.942298 chronyd[1844]: Timezone right/UTC failed leap second check, ignoring Apr 23 23:18:21.942654 chronyd[1844]: Loaded seccomp filter (level 2) Apr 23 23:18:21.948731 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 23 23:18:21.954045 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 23 23:18:21.954483 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 23 23:18:21.955286 systemd[1]: Starting update-engine.service - Update Engine... Apr 23 23:18:21.962332 extend-filesystems[1853]: Found /dev/sda6 Apr 23 23:18:21.966510 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 23 23:18:21.974523 systemd[1]: Started chronyd.service - NTP client/server. Apr 23 23:18:21.987784 extend-filesystems[1853]: Found /dev/sda9 Apr 23 23:18:21.980374 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 23 23:18:22.002046 jq[1876]: true Apr 23 23:18:22.002229 extend-filesystems[1853]: Checking size of /dev/sda9 Apr 23 23:18:21.989453 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 23 23:18:21.992734 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 23 23:18:21.994409 systemd[1]: motdgen.service: Deactivated successfully. Apr 23 23:18:21.994574 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 23 23:18:22.006871 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 23 23:18:22.007008 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 23 23:18:22.029309 extend-filesystems[1853]: Old size kept for /dev/sda9 Apr 23 23:18:22.045188 update_engine[1873]: I20260423 23:18:22.042366 1873 main.cc:92] Flatcar Update Engine starting Apr 23 23:18:22.035252 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 23 23:18:22.045401 jq[1888]: true Apr 23 23:18:22.035890 (ntainerd)[1889]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 23 23:18:22.048434 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 23 23:18:22.048640 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 23 23:18:22.054678 systemd-logind[1871]: New seat seat0. Apr 23 23:18:22.055292 systemd-logind[1871]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 23 23:18:22.057118 systemd[1]: Started systemd-logind.service - User Login Management. Apr 23 23:18:22.083712 tar[1885]: linux-arm64/LICENSE Apr 23 23:18:22.083712 tar[1885]: linux-arm64/helm Apr 23 23:18:22.171280 dbus-daemon[1847]: [system] SELinux support is enabled Apr 23 23:18:22.171739 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 23 23:18:22.180955 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 23 23:18:22.181468 bash[1931]: Updated "/home/core/.ssh/authorized_keys" Apr 23 23:18:22.181847 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 23 23:18:22.187935 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 23 23:18:22.187952 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 23 23:18:22.190370 update_engine[1873]: I20260423 23:18:22.190236 1873 update_check_scheduler.cc:74] Next update check in 9m15s Apr 23 23:18:22.196737 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 23 23:18:22.204811 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Apr 23 23:18:22.205441 dbus-daemon[1847]: [system] Successfully activated service 'org.freedesktop.systemd1' Apr 23 23:18:22.205493 systemd[1]: Started update-engine.service - Update Engine. Apr 23 23:18:22.213133 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 23 23:18:22.255783 sshd_keygen[1881]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 23 23:18:22.326629 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 23 23:18:22.334281 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 23 23:18:22.343542 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Apr 23 23:18:22.367256 systemd[1]: issuegen.service: Deactivated successfully. Apr 23 23:18:22.367692 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 23 23:18:22.377880 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 23 23:18:22.394770 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Apr 23 23:18:22.413980 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 23 23:18:22.420082 coreos-metadata[1846]: Apr 23 23:18:22.420 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Apr 23 23:18:22.423302 coreos-metadata[1846]: Apr 23 23:18:22.423 INFO Fetch successful Apr 23 23:18:22.428782 coreos-metadata[1846]: Apr 23 23:18:22.423 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Apr 23 23:18:22.424377 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 23 23:18:22.433113 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Apr 23 23:18:22.438801 coreos-metadata[1846]: Apr 23 23:18:22.438 INFO Fetch successful Apr 23 23:18:22.439481 coreos-metadata[1846]: Apr 23 23:18:22.439 INFO Fetching http://168.63.129.16/machine/44691888-5549-4a58-8e4d-88a7b73dd4df/e7e2be76%2D1d97%2D4dc8%2Db499%2D1e123c9ad278.%5Fci%2D4459.2.4%2Dn%2D8b75ae451f?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Apr 23 23:18:22.439889 systemd[1]: Reached target getty.target - Login Prompts. Apr 23 23:18:22.447578 coreos-metadata[1846]: Apr 23 23:18:22.447 INFO Fetch successful Apr 23 23:18:22.447578 coreos-metadata[1846]: Apr 23 23:18:22.447 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Apr 23 23:18:22.448974 locksmithd[1950]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 23 23:18:22.453446 coreos-metadata[1846]: Apr 23 23:18:22.453 INFO Fetch successful Apr 23 23:18:22.477442 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 23 23:18:22.483386 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 23 23:18:22.585006 tar[1885]: linux-arm64/README.md Apr 23 23:18:22.598869 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 23 23:18:22.745217 containerd[1889]: time="2026-04-23T23:18:22Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Apr 23 23:18:22.746057 containerd[1889]: time="2026-04-23T23:18:22.745856068Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Apr 23 23:18:22.752164 containerd[1889]: time="2026-04-23T23:18:22.752134100Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="7.08µs" Apr 23 23:18:22.752164 containerd[1889]: time="2026-04-23T23:18:22.752157452Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Apr 23 23:18:22.752222 containerd[1889]: time="2026-04-23T23:18:22.752171556Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Apr 23 23:18:22.752316 containerd[1889]: time="2026-04-23T23:18:22.752295236Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Apr 23 23:18:22.752316 containerd[1889]: time="2026-04-23T23:18:22.752313780Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Apr 23 23:18:22.752362 containerd[1889]: time="2026-04-23T23:18:22.752331228Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752437 containerd[1889]: time="2026-04-23T23:18:22.752372404Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752437 containerd[1889]: time="2026-04-23T23:18:22.752381796Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752546 containerd[1889]: time="2026-04-23T23:18:22.752525068Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752546 containerd[1889]: time="2026-04-23T23:18:22.752541292Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752575 containerd[1889]: time="2026-04-23T23:18:22.752548612Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752575 containerd[1889]: time="2026-04-23T23:18:22.752554212Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752648 containerd[1889]: time="2026-04-23T23:18:22.752635148Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752880 containerd[1889]: time="2026-04-23T23:18:22.752790796Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752880 containerd[1889]: time="2026-04-23T23:18:22.752815084Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Apr 23 23:18:22.752880 containerd[1889]: time="2026-04-23T23:18:22.752820892Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Apr 23 23:18:22.752880 containerd[1889]: time="2026-04-23T23:18:22.752840684Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Apr 23 23:18:22.752994 containerd[1889]: time="2026-04-23T23:18:22.752978404Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Apr 23 23:18:22.753051 containerd[1889]: time="2026-04-23T23:18:22.753032668Z" level=info msg="metadata content store policy set" policy=shared Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769474868Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769515628Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769525188Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769533532Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769540604Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769547212Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769555804Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769562916Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769569556Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769575788Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769581076Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769588356Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Apr 23 23:18:22.769677 containerd[1889]: time="2026-04-23T23:18:22.769686556Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769704236Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769713244Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769720492Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769726716Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769768956Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769776932Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769783188Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769790228Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769796284Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769802436Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769844372Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769853212Z" level=info msg="Start snapshots syncer" Apr 23 23:18:22.770693 containerd[1889]: time="2026-04-23T23:18:22.769866164Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Apr 23 23:18:22.770928 containerd[1889]: time="2026-04-23T23:18:22.770020108Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Apr 23 23:18:22.770928 containerd[1889]: time="2026-04-23T23:18:22.770051284Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770085380Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770165284Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770180396Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770186948Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770198084Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770205684Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770214828Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770221244Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770236388Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770246924Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770255524Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770275068Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770285060Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Apr 23 23:18:22.771007 containerd[1889]: time="2026-04-23T23:18:22.770290036Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770295260Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770299604Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770305740Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770311868Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770322068Z" level=info msg="runtime interface created" Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770325324Z" level=info msg="created NRI interface" Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770330204Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770337484Z" level=info msg="Connect containerd service" Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770349524Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 23 23:18:22.771155 containerd[1889]: time="2026-04-23T23:18:22.770873116Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 23 23:18:22.834560 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:18:22.839554 (kubelet)[2040]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:18:23.136719 kubelet[2040]: E0423 23:18:23.136657 2040 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:18:23.138779 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:18:23.138963 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:18:23.139375 systemd[1]: kubelet.service: Consumed 494ms CPU time, 248.2M memory peak. Apr 23 23:18:23.156430 containerd[1889]: time="2026-04-23T23:18:23.156369748Z" level=info msg="Start subscribing containerd event" Apr 23 23:18:23.156571 containerd[1889]: time="2026-04-23T23:18:23.156502260Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 23 23:18:23.156571 containerd[1889]: time="2026-04-23T23:18:23.156515508Z" level=info msg="Start recovering state" Apr 23 23:18:23.156571 containerd[1889]: time="2026-04-23T23:18:23.156546052Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 23 23:18:23.156944 containerd[1889]: time="2026-04-23T23:18:23.156711332Z" level=info msg="Start event monitor" Apr 23 23:18:23.156944 containerd[1889]: time="2026-04-23T23:18:23.156923956Z" level=info msg="Start cni network conf syncer for default" Apr 23 23:18:23.157077 containerd[1889]: time="2026-04-23T23:18:23.156930724Z" level=info msg="Start streaming server" Apr 23 23:18:23.157077 containerd[1889]: time="2026-04-23T23:18:23.157014708Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Apr 23 23:18:23.157077 containerd[1889]: time="2026-04-23T23:18:23.157025260Z" level=info msg="runtime interface starting up..." Apr 23 23:18:23.157077 containerd[1889]: time="2026-04-23T23:18:23.157032404Z" level=info msg="starting plugins..." Apr 23 23:18:23.157077 containerd[1889]: time="2026-04-23T23:18:23.157054788Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Apr 23 23:18:23.160854 containerd[1889]: time="2026-04-23T23:18:23.157372404Z" level=info msg="containerd successfully booted in 0.412716s" Apr 23 23:18:23.157461 systemd[1]: Started containerd.service - containerd container runtime. Apr 23 23:18:23.164369 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 23 23:18:23.175464 systemd[1]: Startup finished in 1.628s (kernel) + 10.942s (initrd) + 11.037s (userspace) = 23.609s. Apr 23 23:18:23.439796 login[2012]: pam_lastlog(login:session): file /var/log/lastlog is locked/write, retrying Apr 23 23:18:23.441405 login[2013]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:23.446384 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 23 23:18:23.447191 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 23 23:18:23.452666 systemd-logind[1871]: New session 1 of user core. Apr 23 23:18:23.476651 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 23 23:18:23.479695 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 23 23:18:23.487265 (systemd)[2058]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 23 23:18:23.489028 systemd-logind[1871]: New session c1 of user core. Apr 23 23:18:23.605047 systemd[2058]: Queued start job for default target default.target. Apr 23 23:18:23.610236 systemd[2058]: Created slice app.slice - User Application Slice. Apr 23 23:18:23.610547 systemd[2058]: Reached target paths.target - Paths. Apr 23 23:18:23.610650 systemd[2058]: Reached target timers.target - Timers. Apr 23 23:18:23.611550 systemd[2058]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 23 23:18:23.617880 systemd[2058]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 23 23:18:23.617919 systemd[2058]: Reached target sockets.target - Sockets. Apr 23 23:18:23.617946 systemd[2058]: Reached target basic.target - Basic System. Apr 23 23:18:23.617965 systemd[2058]: Reached target default.target - Main User Target. Apr 23 23:18:23.617982 systemd[2058]: Startup finished in 124ms. Apr 23 23:18:23.618068 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 23 23:18:23.623817 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 23 23:18:24.032614 waagent[2009]: 2026-04-23T23:18:24.032095Z INFO Daemon Daemon Azure Linux Agent Version: 2.12.0.4 Apr 23 23:18:24.036298 waagent[2009]: 2026-04-23T23:18:24.036257Z INFO Daemon Daemon OS: flatcar 4459.2.4 Apr 23 23:18:24.039559 waagent[2009]: 2026-04-23T23:18:24.039529Z INFO Daemon Daemon Python: 3.11.13 Apr 23 23:18:24.042875 waagent[2009]: 2026-04-23T23:18:24.042839Z INFO Daemon Daemon Run daemon Apr 23 23:18:24.045961 waagent[2009]: 2026-04-23T23:18:24.045923Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4459.2.4' Apr 23 23:18:24.052546 waagent[2009]: 2026-04-23T23:18:24.052507Z INFO Daemon Daemon Using waagent for provisioning Apr 23 23:18:24.056321 waagent[2009]: 2026-04-23T23:18:24.056281Z INFO Daemon Daemon Activate resource disk Apr 23 23:18:24.060282 waagent[2009]: 2026-04-23T23:18:24.060251Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Apr 23 23:18:24.068418 waagent[2009]: 2026-04-23T23:18:24.068381Z INFO Daemon Daemon Found device: None Apr 23 23:18:24.071907 waagent[2009]: 2026-04-23T23:18:24.071876Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Apr 23 23:18:24.078131 waagent[2009]: 2026-04-23T23:18:24.078105Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Apr 23 23:18:24.086402 waagent[2009]: 2026-04-23T23:18:24.086366Z INFO Daemon Daemon Clean protocol and wireserver endpoint Apr 23 23:18:24.090548 waagent[2009]: 2026-04-23T23:18:24.090519Z INFO Daemon Daemon Running default provisioning handler Apr 23 23:18:24.098650 waagent[2009]: 2026-04-23T23:18:24.098590Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Apr 23 23:18:24.108473 waagent[2009]: 2026-04-23T23:18:24.108433Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Apr 23 23:18:24.115391 waagent[2009]: 2026-04-23T23:18:24.115358Z INFO Daemon Daemon cloud-init is enabled: False Apr 23 23:18:24.119058 waagent[2009]: 2026-04-23T23:18:24.119032Z INFO Daemon Daemon Copying ovf-env.xml Apr 23 23:18:24.195641 waagent[2009]: 2026-04-23T23:18:24.195493Z INFO Daemon Daemon Successfully mounted dvd Apr 23 23:18:24.220516 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Apr 23 23:18:24.222565 waagent[2009]: 2026-04-23T23:18:24.222514Z INFO Daemon Daemon Detect protocol endpoint Apr 23 23:18:24.226110 waagent[2009]: 2026-04-23T23:18:24.226076Z INFO Daemon Daemon Clean protocol and wireserver endpoint Apr 23 23:18:24.230207 waagent[2009]: 2026-04-23T23:18:24.230179Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Apr 23 23:18:24.235132 waagent[2009]: 2026-04-23T23:18:24.235107Z INFO Daemon Daemon Test for route to 168.63.129.16 Apr 23 23:18:24.238959 waagent[2009]: 2026-04-23T23:18:24.238927Z INFO Daemon Daemon Route to 168.63.129.16 exists Apr 23 23:18:24.242624 waagent[2009]: 2026-04-23T23:18:24.242589Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Apr 23 23:18:24.285598 waagent[2009]: 2026-04-23T23:18:24.285516Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Apr 23 23:18:24.290227 waagent[2009]: 2026-04-23T23:18:24.290207Z INFO Daemon Daemon Wire protocol version:2012-11-30 Apr 23 23:18:24.293888 waagent[2009]: 2026-04-23T23:18:24.293863Z INFO Daemon Daemon Server preferred version:2015-04-05 Apr 23 23:18:24.427659 waagent[2009]: 2026-04-23T23:18:24.427584Z INFO Daemon Daemon Initializing goal state during protocol detection Apr 23 23:18:24.432127 waagent[2009]: 2026-04-23T23:18:24.432096Z INFO Daemon Daemon Forcing an update of the goal state. Apr 23 23:18:24.440136 login[2012]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:24.440489 waagent[2009]: 2026-04-23T23:18:24.440327Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Apr 23 23:18:24.448761 systemd-logind[1871]: New session 2 of user core. Apr 23 23:18:24.454862 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 23 23:18:24.458182 waagent[2009]: 2026-04-23T23:18:24.458146Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.181 Apr 23 23:18:24.462844 waagent[2009]: 2026-04-23T23:18:24.462803Z INFO Daemon Apr 23 23:18:24.464874 waagent[2009]: 2026-04-23T23:18:24.464843Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: af872af2-6bd2-4659-9f4c-92834c748e2a eTag: 15372486120474941090 source: Fabric] Apr 23 23:18:24.473677 waagent[2009]: 2026-04-23T23:18:24.473640Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Apr 23 23:18:24.478570 waagent[2009]: 2026-04-23T23:18:24.478533Z INFO Daemon Apr 23 23:18:24.480641 waagent[2009]: 2026-04-23T23:18:24.480612Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Apr 23 23:18:24.488295 waagent[2009]: 2026-04-23T23:18:24.488268Z INFO Daemon Daemon Downloading artifacts profile blob Apr 23 23:18:24.594865 waagent[2009]: 2026-04-23T23:18:24.594785Z INFO Daemon Downloaded certificate {'thumbprint': '3FF5E2F9E02E89D9984EEFDE699762914A5EF9D7', 'hasPrivateKey': True} Apr 23 23:18:24.601798 waagent[2009]: 2026-04-23T23:18:24.601763Z INFO Daemon Fetch goal state completed Apr 23 23:18:24.636164 waagent[2009]: 2026-04-23T23:18:24.636131Z INFO Daemon Daemon Starting provisioning Apr 23 23:18:24.639909 waagent[2009]: 2026-04-23T23:18:24.639879Z INFO Daemon Daemon Handle ovf-env.xml. Apr 23 23:18:24.643699 waagent[2009]: 2026-04-23T23:18:24.643676Z INFO Daemon Daemon Set hostname [ci-4459.2.4-n-8b75ae451f] Apr 23 23:18:24.649221 waagent[2009]: 2026-04-23T23:18:24.649178Z INFO Daemon Daemon Publish hostname [ci-4459.2.4-n-8b75ae451f] Apr 23 23:18:24.653714 waagent[2009]: 2026-04-23T23:18:24.653680Z INFO Daemon Daemon Examine /proc/net/route for primary interface Apr 23 23:18:24.658554 waagent[2009]: 2026-04-23T23:18:24.658523Z INFO Daemon Daemon Primary interface is [eth0] Apr 23 23:18:24.667317 systemd-networkd[1474]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:18:24.667322 systemd-networkd[1474]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 23 23:18:24.667356 systemd-networkd[1474]: eth0: DHCP lease lost Apr 23 23:18:24.668116 waagent[2009]: 2026-04-23T23:18:24.668072Z INFO Daemon Daemon Create user account if not exists Apr 23 23:18:24.672218 waagent[2009]: 2026-04-23T23:18:24.672186Z INFO Daemon Daemon User core already exists, skip useradd Apr 23 23:18:24.676433 waagent[2009]: 2026-04-23T23:18:24.676407Z INFO Daemon Daemon Configure sudoer Apr 23 23:18:24.679749 waagent[2009]: 2026-04-23T23:18:24.679705Z INFO Daemon Daemon Configure sshd Apr 23 23:18:24.682923 waagent[2009]: 2026-04-23T23:18:24.682889Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Apr 23 23:18:24.692114 waagent[2009]: 2026-04-23T23:18:24.692085Z INFO Daemon Daemon Deploy ssh public key. Apr 23 23:18:24.720634 systemd-networkd[1474]: eth0: DHCPv4 address 10.0.0.13/24, gateway 10.0.0.1 acquired from 168.63.129.16 Apr 23 23:18:25.783724 waagent[2009]: 2026-04-23T23:18:25.783667Z INFO Daemon Daemon Provisioning complete Apr 23 23:18:25.795231 waagent[2009]: 2026-04-23T23:18:25.795194Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Apr 23 23:18:25.799813 waagent[2009]: 2026-04-23T23:18:25.799781Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Apr 23 23:18:25.806516 waagent[2009]: 2026-04-23T23:18:25.806490Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.12.0.4 is the most current agent Apr 23 23:18:25.904405 waagent[2108]: 2026-04-23T23:18:25.904355Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.12.0.4) Apr 23 23:18:25.905753 waagent[2108]: 2026-04-23T23:18:25.904703Z INFO ExtHandler ExtHandler OS: flatcar 4459.2.4 Apr 23 23:18:25.905753 waagent[2108]: 2026-04-23T23:18:25.904751Z INFO ExtHandler ExtHandler Python: 3.11.13 Apr 23 23:18:25.905753 waagent[2108]: 2026-04-23T23:18:25.904785Z INFO ExtHandler ExtHandler CPU Arch: aarch64 Apr 23 23:18:25.943997 waagent[2108]: 2026-04-23T23:18:25.943961Z INFO ExtHandler ExtHandler Distro: flatcar-4459.2.4; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.13; Arch: aarch64; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.22.0; Apr 23 23:18:25.944196 waagent[2108]: 2026-04-23T23:18:25.944167Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 23 23:18:25.944309 waagent[2108]: 2026-04-23T23:18:25.944285Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 23 23:18:25.949221 waagent[2108]: 2026-04-23T23:18:25.949177Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Apr 23 23:18:25.953397 waagent[2108]: 2026-04-23T23:18:25.953369Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.181 Apr 23 23:18:25.953843 waagent[2108]: 2026-04-23T23:18:25.953810Z INFO ExtHandler Apr 23 23:18:25.953964 waagent[2108]: 2026-04-23T23:18:25.953940Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 89686298-2372-42b8-88e0-f2243479efa2 eTag: 15372486120474941090 source: Fabric] Apr 23 23:18:25.954284 waagent[2108]: 2026-04-23T23:18:25.954255Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Apr 23 23:18:25.954806 waagent[2108]: 2026-04-23T23:18:25.954773Z INFO ExtHandler Apr 23 23:18:25.954924 waagent[2108]: 2026-04-23T23:18:25.954901Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Apr 23 23:18:25.957426 waagent[2108]: 2026-04-23T23:18:25.957399Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Apr 23 23:18:26.006464 waagent[2108]: 2026-04-23T23:18:26.006427Z INFO ExtHandler Downloaded certificate {'thumbprint': '3FF5E2F9E02E89D9984EEFDE699762914A5EF9D7', 'hasPrivateKey': True} Apr 23 23:18:26.007023 waagent[2108]: 2026-04-23T23:18:26.006987Z INFO ExtHandler Fetch goal state completed Apr 23 23:18:26.016939 waagent[2108]: 2026-04-23T23:18:26.016908Z INFO ExtHandler ExtHandler OpenSSL version: OpenSSL 3.4.4 27 Jan 2026 (Library: OpenSSL 3.4.4 27 Jan 2026) Apr 23 23:18:26.020073 waagent[2108]: 2026-04-23T23:18:26.020037Z INFO ExtHandler ExtHandler WALinuxAgent-2.12.0.4 running as process 2108 Apr 23 23:18:26.020243 waagent[2108]: 2026-04-23T23:18:26.020215Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Apr 23 23:18:26.020560 waagent[2108]: 2026-04-23T23:18:26.020529Z INFO ExtHandler ExtHandler ******** AutoUpdate.UpdateToLatestVersion is set to False, not processing the operation ******** Apr 23 23:18:26.021789 waagent[2108]: 2026-04-23T23:18:26.021754Z INFO ExtHandler ExtHandler [CGI] Cgroup monitoring is not supported on ['flatcar', '4459.2.4', '', 'Flatcar Container Linux by Kinvolk'] Apr 23 23:18:26.022192 waagent[2108]: 2026-04-23T23:18:26.022160Z INFO ExtHandler ExtHandler [CGI] Agent will reset the quotas in case distro: ['flatcar', '4459.2.4', '', 'Flatcar Container Linux by Kinvolk'] went from supported to unsupported Apr 23 23:18:26.022366 waagent[2108]: 2026-04-23T23:18:26.022340Z INFO ExtHandler ExtHandler [CGI] Agent cgroups enabled: False Apr 23 23:18:26.022898 waagent[2108]: 2026-04-23T23:18:26.022866Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Apr 23 23:18:26.087550 waagent[2108]: 2026-04-23T23:18:26.087211Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Apr 23 23:18:26.087550 waagent[2108]: 2026-04-23T23:18:26.087342Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Apr 23 23:18:26.091622 waagent[2108]: 2026-04-23T23:18:26.091570Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Apr 23 23:18:26.095850 systemd[1]: Reload requested from client PID 2123 ('systemctl') (unit waagent.service)... Apr 23 23:18:26.096032 systemd[1]: Reloading... Apr 23 23:18:26.163642 zram_generator::config[2177]: No configuration found. Apr 23 23:18:26.294059 systemd[1]: Reloading finished in 197 ms. Apr 23 23:18:26.313623 waagent[2108]: 2026-04-23T23:18:26.312209Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Apr 23 23:18:26.313623 waagent[2108]: 2026-04-23T23:18:26.312329Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Apr 23 23:18:26.754309 waagent[2108]: 2026-04-23T23:18:26.753564Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Apr 23 23:18:26.754309 waagent[2108]: 2026-04-23T23:18:26.753876Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: 1. configuration enabled [True], 2. cgroups v1 enabled [False] OR cgroups v2 is in use and v2 resource limiting configuration enabled [False], 3. python supported: [True] Apr 23 23:18:26.754535 waagent[2108]: 2026-04-23T23:18:26.754492Z INFO ExtHandler ExtHandler Starting env monitor service. Apr 23 23:18:26.754651 waagent[2108]: 2026-04-23T23:18:26.754613Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 23 23:18:26.755018 waagent[2108]: 2026-04-23T23:18:26.754981Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Apr 23 23:18:26.755073 waagent[2108]: 2026-04-23T23:18:26.755041Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 23 23:18:26.755258 waagent[2108]: 2026-04-23T23:18:26.755236Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 23 23:18:26.755415 waagent[2108]: 2026-04-23T23:18:26.755384Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Apr 23 23:18:26.755738 waagent[2108]: 2026-04-23T23:18:26.755695Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Apr 23 23:18:26.755871 waagent[2108]: 2026-04-23T23:18:26.755811Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Apr 23 23:18:26.755918 waagent[2108]: 2026-04-23T23:18:26.755893Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 23 23:18:26.756031 waagent[2108]: 2026-04-23T23:18:26.756001Z INFO EnvHandler ExtHandler Configure routes Apr 23 23:18:26.756072 waagent[2108]: 2026-04-23T23:18:26.756054Z INFO EnvHandler ExtHandler Gateway:None Apr 23 23:18:26.756092 waagent[2108]: 2026-04-23T23:18:26.756083Z INFO EnvHandler ExtHandler Routes:None Apr 23 23:18:26.756440 waagent[2108]: 2026-04-23T23:18:26.756403Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Apr 23 23:18:26.756489 waagent[2108]: 2026-04-23T23:18:26.756478Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Apr 23 23:18:26.756620 waagent[2108]: 2026-04-23T23:18:26.756555Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Apr 23 23:18:26.756769 waagent[2108]: 2026-04-23T23:18:26.756737Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Apr 23 23:18:26.756769 waagent[2108]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Apr 23 23:18:26.756769 waagent[2108]: eth0 00000000 0100000A 0003 0 0 1024 00000000 0 0 0 Apr 23 23:18:26.756769 waagent[2108]: eth0 0000000A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Apr 23 23:18:26.756769 waagent[2108]: eth0 0100000A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Apr 23 23:18:26.756769 waagent[2108]: eth0 10813FA8 0100000A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 23 23:18:26.756769 waagent[2108]: eth0 FEA9FEA9 0100000A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 23 23:18:26.761497 waagent[2108]: 2026-04-23T23:18:26.761446Z INFO ExtHandler ExtHandler Apr 23 23:18:26.761733 waagent[2108]: 2026-04-23T23:18:26.761697Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: f89c8504-5694-48b3-b59a-03fc4d9aee26 correlation 8a31020a-6dec-40d0-9dc3-dee17a0a428b created: 2026-04-23T23:17:32.161392Z] Apr 23 23:18:26.762729 waagent[2108]: 2026-04-23T23:18:26.762687Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Apr 23 23:18:26.763621 waagent[2108]: 2026-04-23T23:18:26.763317Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 1 ms] Apr 23 23:18:26.785202 waagent[2108]: 2026-04-23T23:18:26.785155Z WARNING ExtHandler ExtHandler Failed to get firewall packets: 'iptables -w -t security -L OUTPUT --zero OUTPUT -nxv' failed: 2 (iptables v1.8.11 (nf_tables): Illegal option `--numeric' with this command Apr 23 23:18:26.785202 waagent[2108]: Try `iptables -h' or 'iptables --help' for more information.) Apr 23 23:18:26.785490 waagent[2108]: 2026-04-23T23:18:26.785455Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.12.0.4 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 63541EFB-CEDC-45F2-AE9C-2C699CAECE2D;DroppedPackets: -1;UpdateGSErrors: 0;AutoUpdate: 0;UpdateMode: SelfUpdate;] Apr 23 23:18:26.833470 waagent[2108]: 2026-04-23T23:18:26.833182Z INFO MonitorHandler ExtHandler Network interfaces: Apr 23 23:18:26.833470 waagent[2108]: Executing ['ip', '-a', '-o', 'link']: Apr 23 23:18:26.833470 waagent[2108]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Apr 23 23:18:26.833470 waagent[2108]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:f5:92:60 brd ff:ff:ff:ff:ff:ff Apr 23 23:18:26.833470 waagent[2108]: 3: enP30347s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:f5:92:60 brd ff:ff:ff:ff:ff:ff\ altname enP30347p0s2 Apr 23 23:18:26.833470 waagent[2108]: Executing ['ip', '-4', '-a', '-o', 'address']: Apr 23 23:18:26.833470 waagent[2108]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Apr 23 23:18:26.833470 waagent[2108]: 2: eth0 inet 10.0.0.13/24 metric 1024 brd 10.0.0.255 scope global eth0\ valid_lft forever preferred_lft forever Apr 23 23:18:26.833470 waagent[2108]: Executing ['ip', '-6', '-a', '-o', 'address']: Apr 23 23:18:26.833470 waagent[2108]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Apr 23 23:18:26.833470 waagent[2108]: 2: eth0 inet6 fe80::20d:3aff:fef5:9260/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Apr 23 23:18:26.866527 waagent[2108]: 2026-04-23T23:18:26.866495Z INFO EnvHandler ExtHandler Created firewall rules for the Azure Fabric: Apr 23 23:18:26.866527 waagent[2108]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Apr 23 23:18:26.866527 waagent[2108]: pkts bytes target prot opt in out source destination Apr 23 23:18:26.866527 waagent[2108]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Apr 23 23:18:26.866527 waagent[2108]: pkts bytes target prot opt in out source destination Apr 23 23:18:26.866527 waagent[2108]: Chain OUTPUT (policy ACCEPT 3 packets, 349 bytes) Apr 23 23:18:26.866527 waagent[2108]: pkts bytes target prot opt in out source destination Apr 23 23:18:26.866527 waagent[2108]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Apr 23 23:18:26.866527 waagent[2108]: 1 52 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Apr 23 23:18:26.866527 waagent[2108]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Apr 23 23:18:26.868989 waagent[2108]: 2026-04-23T23:18:26.868958Z INFO EnvHandler ExtHandler Current Firewall rules: Apr 23 23:18:26.868989 waagent[2108]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Apr 23 23:18:26.868989 waagent[2108]: pkts bytes target prot opt in out source destination Apr 23 23:18:26.868989 waagent[2108]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Apr 23 23:18:26.868989 waagent[2108]: pkts bytes target prot opt in out source destination Apr 23 23:18:26.868989 waagent[2108]: Chain OUTPUT (policy ACCEPT 3 packets, 349 bytes) Apr 23 23:18:26.868989 waagent[2108]: pkts bytes target prot opt in out source destination Apr 23 23:18:26.868989 waagent[2108]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Apr 23 23:18:26.868989 waagent[2108]: 1 52 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Apr 23 23:18:26.868989 waagent[2108]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Apr 23 23:18:26.869373 waagent[2108]: 2026-04-23T23:18:26.869350Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Apr 23 23:18:33.377401 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 23 23:18:33.378679 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:18:33.476202 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:18:33.484800 (kubelet)[2257]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:18:33.593000 kubelet[2257]: E0423 23:18:33.592939 2257 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:18:33.595580 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:18:33.595799 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:18:33.596252 systemd[1]: kubelet.service: Consumed 108ms CPU time, 107.2M memory peak. Apr 23 23:18:43.626943 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 23 23:18:43.628166 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:18:43.727114 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:18:43.733831 (kubelet)[2272]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:18:43.832450 kubelet[2272]: E0423 23:18:43.832395 2272 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:18:43.834497 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:18:43.834679 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:18:43.834932 systemd[1]: kubelet.service: Consumed 102ms CPU time, 105.1M memory peak. Apr 23 23:18:45.733684 chronyd[1844]: Selected source PHC0 Apr 23 23:18:51.096710 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 23 23:18:51.097968 systemd[1]: Started sshd@0-10.0.0.13:22-50.85.169.122:36008.service - OpenSSH per-connection server daemon (50.85.169.122:36008). Apr 23 23:18:52.012815 sshd[2280]: Accepted publickey for core from 50.85.169.122 port 36008 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:18:52.014217 sshd-session[2280]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:52.017655 systemd-logind[1871]: New session 3 of user core. Apr 23 23:18:52.027884 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 23 23:18:52.603515 systemd[1]: Started sshd@1-10.0.0.13:22-50.85.169.122:36020.service - OpenSSH per-connection server daemon (50.85.169.122:36020). Apr 23 23:18:53.380511 sshd[2286]: Accepted publickey for core from 50.85.169.122 port 36020 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:18:53.381206 sshd-session[2286]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:53.384737 systemd-logind[1871]: New session 4 of user core. Apr 23 23:18:53.392711 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 23 23:18:53.829659 sshd[2289]: Connection closed by 50.85.169.122 port 36020 Apr 23 23:18:53.829497 sshd-session[2286]: pam_unix(sshd:session): session closed for user core Apr 23 23:18:53.833411 systemd[1]: sshd@1-10.0.0.13:22-50.85.169.122:36020.service: Deactivated successfully. Apr 23 23:18:53.835220 systemd[1]: session-4.scope: Deactivated successfully. Apr 23 23:18:53.836216 systemd-logind[1871]: Session 4 logged out. Waiting for processes to exit. Apr 23 23:18:53.837566 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Apr 23 23:18:53.839473 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:18:53.840066 systemd-logind[1871]: Removed session 4. Apr 23 23:18:53.939103 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:18:53.946837 (kubelet)[2302]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:18:53.972206 kubelet[2302]: E0423 23:18:53.972159 2302 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:18:53.974994 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:18:53.975088 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:18:53.975704 systemd[1]: kubelet.service: Consumed 101ms CPU time, 104.6M memory peak. Apr 23 23:18:53.977896 systemd[1]: Started sshd@2-10.0.0.13:22-50.85.169.122:36024.service - OpenSSH per-connection server daemon (50.85.169.122:36024). Apr 23 23:18:54.712272 sshd[2309]: Accepted publickey for core from 50.85.169.122 port 36024 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:18:54.712963 sshd-session[2309]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:54.716640 systemd-logind[1871]: New session 5 of user core. Apr 23 23:18:54.726689 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 23 23:18:55.132448 sshd[2312]: Connection closed by 50.85.169.122 port 36024 Apr 23 23:18:55.132370 sshd-session[2309]: pam_unix(sshd:session): session closed for user core Apr 23 23:18:55.135509 systemd-logind[1871]: Session 5 logged out. Waiting for processes to exit. Apr 23 23:18:55.136049 systemd[1]: sshd@2-10.0.0.13:22-50.85.169.122:36024.service: Deactivated successfully. Apr 23 23:18:55.137576 systemd[1]: session-5.scope: Deactivated successfully. Apr 23 23:18:55.139045 systemd-logind[1871]: Removed session 5. Apr 23 23:18:55.293675 systemd[1]: Started sshd@3-10.0.0.13:22-50.85.169.122:36032.service - OpenSSH per-connection server daemon (50.85.169.122:36032). Apr 23 23:18:56.061625 sshd[2318]: Accepted publickey for core from 50.85.169.122 port 36032 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:18:56.062537 sshd-session[2318]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:56.065986 systemd-logind[1871]: New session 6 of user core. Apr 23 23:18:56.071731 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 23 23:18:56.506633 sshd[2321]: Connection closed by 50.85.169.122 port 36032 Apr 23 23:18:56.507177 sshd-session[2318]: pam_unix(sshd:session): session closed for user core Apr 23 23:18:56.510182 systemd[1]: sshd@3-10.0.0.13:22-50.85.169.122:36032.service: Deactivated successfully. Apr 23 23:18:56.511869 systemd[1]: session-6.scope: Deactivated successfully. Apr 23 23:18:56.513133 systemd-logind[1871]: Session 6 logged out. Waiting for processes to exit. Apr 23 23:18:56.514213 systemd-logind[1871]: Removed session 6. Apr 23 23:18:56.665064 systemd[1]: Started sshd@4-10.0.0.13:22-50.85.169.122:36038.service - OpenSSH per-connection server daemon (50.85.169.122:36038). Apr 23 23:18:57.433279 sshd[2327]: Accepted publickey for core from 50.85.169.122 port 36038 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:18:57.434311 sshd-session[2327]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:57.437884 systemd-logind[1871]: New session 7 of user core. Apr 23 23:18:57.444720 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 23 23:18:57.911387 sudo[2331]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 23 23:18:57.911631 sudo[2331]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:18:57.942866 sudo[2331]: pam_unix(sudo:session): session closed for user root Apr 23 23:18:58.090472 sshd[2330]: Connection closed by 50.85.169.122 port 36038 Apr 23 23:18:58.091140 sshd-session[2327]: pam_unix(sshd:session): session closed for user core Apr 23 23:18:58.094554 systemd[1]: sshd@4-10.0.0.13:22-50.85.169.122:36038.service: Deactivated successfully. Apr 23 23:18:58.096142 systemd[1]: session-7.scope: Deactivated successfully. Apr 23 23:18:58.096757 systemd-logind[1871]: Session 7 logged out. Waiting for processes to exit. Apr 23 23:18:58.098138 systemd-logind[1871]: Removed session 7. Apr 23 23:18:58.241798 systemd[1]: Started sshd@5-10.0.0.13:22-50.85.169.122:36054.service - OpenSSH per-connection server daemon (50.85.169.122:36054). Apr 23 23:18:58.985511 sshd[2337]: Accepted publickey for core from 50.85.169.122 port 36054 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:18:58.986520 sshd-session[2337]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:18:58.990953 systemd-logind[1871]: New session 8 of user core. Apr 23 23:18:58.997728 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 23 23:18:59.273799 sudo[2342]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 23 23:18:59.273998 sudo[2342]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:18:59.276881 sudo[2342]: pam_unix(sudo:session): session closed for user root Apr 23 23:18:59.280164 sudo[2341]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Apr 23 23:18:59.280565 sudo[2341]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:18:59.286772 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 23 23:18:59.315253 augenrules[2364]: No rules Apr 23 23:18:59.316279 systemd[1]: audit-rules.service: Deactivated successfully. Apr 23 23:18:59.316426 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 23 23:18:59.317293 sudo[2341]: pam_unix(sudo:session): session closed for user root Apr 23 23:18:59.460170 sshd[2340]: Connection closed by 50.85.169.122 port 36054 Apr 23 23:18:59.460750 sshd-session[2337]: pam_unix(sshd:session): session closed for user core Apr 23 23:18:59.463172 systemd[1]: sshd@5-10.0.0.13:22-50.85.169.122:36054.service: Deactivated successfully. Apr 23 23:18:59.464737 systemd[1]: session-8.scope: Deactivated successfully. Apr 23 23:18:59.465845 systemd-logind[1871]: Session 8 logged out. Waiting for processes to exit. Apr 23 23:18:59.467530 systemd-logind[1871]: Removed session 8. Apr 23 23:18:59.622885 systemd[1]: Started sshd@6-10.0.0.13:22-50.85.169.122:53780.service - OpenSSH per-connection server daemon (50.85.169.122:53780). Apr 23 23:19:00.393337 sshd[2373]: Accepted publickey for core from 50.85.169.122 port 53780 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:19:00.393999 sshd-session[2373]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:19:00.397358 systemd-logind[1871]: New session 9 of user core. Apr 23 23:19:00.403692 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 23 23:19:00.692255 sudo[2377]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 23 23:19:00.692491 sudo[2377]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:19:02.241765 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 23 23:19:02.250970 (dockerd)[2395]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 23 23:19:03.244634 dockerd[2395]: time="2026-04-23T23:19:03.244530417Z" level=info msg="Starting up" Apr 23 23:19:03.245220 dockerd[2395]: time="2026-04-23T23:19:03.245196019Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Apr 23 23:19:03.253950 dockerd[2395]: time="2026-04-23T23:19:03.253919154Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Apr 23 23:19:03.339109 dockerd[2395]: time="2026-04-23T23:19:03.339078398Z" level=info msg="Loading containers: start." Apr 23 23:19:03.368624 kernel: Initializing XFRM netlink socket Apr 23 23:19:03.692733 systemd-networkd[1474]: docker0: Link UP Apr 23 23:19:03.712017 dockerd[2395]: time="2026-04-23T23:19:03.711984273Z" level=info msg="Loading containers: done." Apr 23 23:19:03.721924 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1542349018-merged.mount: Deactivated successfully. Apr 23 23:19:03.730947 dockerd[2395]: time="2026-04-23T23:19:03.730911433Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 23 23:19:03.731011 dockerd[2395]: time="2026-04-23T23:19:03.730977188Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Apr 23 23:19:03.731062 dockerd[2395]: time="2026-04-23T23:19:03.731046519Z" level=info msg="Initializing buildkit" Apr 23 23:19:03.778232 dockerd[2395]: time="2026-04-23T23:19:03.778162091Z" level=info msg="Completed buildkit initialization" Apr 23 23:19:03.782872 dockerd[2395]: time="2026-04-23T23:19:03.782613034Z" level=info msg="Daemon has completed initialization" Apr 23 23:19:03.782872 dockerd[2395]: time="2026-04-23T23:19:03.782648707Z" level=info msg="API listen on /run/docker.sock" Apr 23 23:19:03.782677 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 23 23:19:04.126802 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Apr 23 23:19:04.128750 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:19:04.166414 containerd[1889]: time="2026-04-23T23:19:04.166077052Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.7\"" Apr 23 23:19:04.336928 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:19:04.343793 (kubelet)[2610]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:19:04.367035 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:19:04.628298 kubelet[2610]: E0423 23:19:04.365845 2610 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:19:04.367118 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:19:04.367475 systemd[1]: kubelet.service: Consumed 100ms CPU time, 106.8M memory peak. Apr 23 23:19:05.319195 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Apr 23 23:19:05.507424 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1496047741.mount: Deactivated successfully. Apr 23 23:19:06.775623 containerd[1889]: time="2026-04-23T23:19:06.775416853Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:06.778611 containerd[1889]: time="2026-04-23T23:19:06.778426171Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.7: active requests=0, bytes read=24193768" Apr 23 23:19:06.782370 containerd[1889]: time="2026-04-23T23:19:06.782338013Z" level=info msg="ImageCreate event name:\"sha256:bf3fdee5548e267fd53c67a79d712e896d47f48203512415518d59da7f985228\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:06.786616 containerd[1889]: time="2026-04-23T23:19:06.786435142Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:b96b8464d152a24c81d7f0435fd2198f8486970cd26a9e0e9c20826c73d1441c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:06.787357 containerd[1889]: time="2026-04-23T23:19:06.787336617Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.7\" with image id \"sha256:bf3fdee5548e267fd53c67a79d712e896d47f48203512415518d59da7f985228\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:b96b8464d152a24c81d7f0435fd2198f8486970cd26a9e0e9c20826c73d1441c\", size \"24190367\" in 2.621224756s" Apr 23 23:19:06.787451 containerd[1889]: time="2026-04-23T23:19:06.787438621Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.7\" returns image reference \"sha256:bf3fdee5548e267fd53c67a79d712e896d47f48203512415518d59da7f985228\"" Apr 23 23:19:06.787938 containerd[1889]: time="2026-04-23T23:19:06.787914408Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.7\"" Apr 23 23:19:07.567909 update_engine[1873]: I20260423 23:19:07.567847 1873 update_attempter.cc:509] Updating boot flags... Apr 23 23:19:08.012771 containerd[1889]: time="2026-04-23T23:19:08.012724085Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:08.015522 containerd[1889]: time="2026-04-23T23:19:08.015500802Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.7: active requests=0, bytes read=18901444" Apr 23 23:19:08.018872 containerd[1889]: time="2026-04-23T23:19:08.018852190Z" level=info msg="ImageCreate event name:\"sha256:161b12aee2701d72b2e8a7d114f5f83122603d8c5d1d3cd7f72aa6fac5d9524c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:08.025440 containerd[1889]: time="2026-04-23T23:19:08.025416752Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7d759bdc4fef10a3fc1ad60ce9439d58e1a4df7ebb22751f7cc0201ce55f280b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:08.026080 containerd[1889]: time="2026-04-23T23:19:08.025921900Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.7\" with image id \"sha256:161b12aee2701d72b2e8a7d114f5f83122603d8c5d1d3cd7f72aa6fac5d9524c\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7d759bdc4fef10a3fc1ad60ce9439d58e1a4df7ebb22751f7cc0201ce55f280b\", size \"20408083\" in 1.237983627s" Apr 23 23:19:08.026080 containerd[1889]: time="2026-04-23T23:19:08.025950429Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.7\" returns image reference \"sha256:161b12aee2701d72b2e8a7d114f5f83122603d8c5d1d3cd7f72aa6fac5d9524c\"" Apr 23 23:19:08.026395 containerd[1889]: time="2026-04-23T23:19:08.026378022Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.7\"" Apr 23 23:19:09.211400 containerd[1889]: time="2026-04-23T23:19:09.211347781Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:09.214197 containerd[1889]: time="2026-04-23T23:19:09.213980316Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.7: active requests=0, bytes read=14047945" Apr 23 23:19:09.217742 containerd[1889]: time="2026-04-23T23:19:09.217718647Z" level=info msg="ImageCreate event name:\"sha256:85bc0b83d6779f309f0f2d8724ee225e2a061dc60b1b127f8a9b8843bad36e14\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:09.222186 containerd[1889]: time="2026-04-23T23:19:09.222160190Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:4ab32f707ff84beaac431797999707757b885196b0b9a52d29cb67f95efce7c1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:09.222693 containerd[1889]: time="2026-04-23T23:19:09.222670922Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.7\" with image id \"sha256:85bc0b83d6779f309f0f2d8724ee225e2a061dc60b1b127f8a9b8843bad36e14\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:4ab32f707ff84beaac431797999707757b885196b0b9a52d29cb67f95efce7c1\", size \"15554602\" in 1.195904965s" Apr 23 23:19:09.222778 containerd[1889]: time="2026-04-23T23:19:09.222763653Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.7\" returns image reference \"sha256:85bc0b83d6779f309f0f2d8724ee225e2a061dc60b1b127f8a9b8843bad36e14\"" Apr 23 23:19:09.223314 containerd[1889]: time="2026-04-23T23:19:09.223289474Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.7\"" Apr 23 23:19:10.636346 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3597491600.mount: Deactivated successfully. Apr 23 23:19:10.845497 containerd[1889]: time="2026-04-23T23:19:10.845443151Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:10.849108 containerd[1889]: time="2026-04-23T23:19:10.849081542Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.7: active requests=0, bytes read=22606286" Apr 23 23:19:10.852214 containerd[1889]: time="2026-04-23T23:19:10.852188977Z" level=info msg="ImageCreate event name:\"sha256:c63683691df94ddfb3e7b1449f68fd9df087b1bda7cdecd1e9292214f6adc745\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:10.856544 containerd[1889]: time="2026-04-23T23:19:10.856514466Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:062519bc0a14769e2f98c6bdff7816a17e6252de3f3c9cb102e6be33fe38d9e2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:10.857120 containerd[1889]: time="2026-04-23T23:19:10.856785652Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.7\" with image id \"sha256:c63683691df94ddfb3e7b1449f68fd9df087b1bda7cdecd1e9292214f6adc745\", repo tag \"registry.k8s.io/kube-proxy:v1.34.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:062519bc0a14769e2f98c6bdff7816a17e6252de3f3c9cb102e6be33fe38d9e2\", size \"22605305\" in 1.633469089s" Apr 23 23:19:10.857120 containerd[1889]: time="2026-04-23T23:19:10.856808709Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.7\" returns image reference \"sha256:c63683691df94ddfb3e7b1449f68fd9df087b1bda7cdecd1e9292214f6adc745\"" Apr 23 23:19:10.857196 containerd[1889]: time="2026-04-23T23:19:10.857178747Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Apr 23 23:19:11.619765 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount197288448.mount: Deactivated successfully. Apr 23 23:19:12.638358 containerd[1889]: time="2026-04-23T23:19:12.638289616Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:12.641710 containerd[1889]: time="2026-04-23T23:19:12.641678318Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=20395406" Apr 23 23:19:12.644940 containerd[1889]: time="2026-04-23T23:19:12.644918598Z" level=info msg="ImageCreate event name:\"sha256:138784d87c9c50f8e59412544da4cf4928d61ccbaf93b9f5898a3ba406871bfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:12.649643 containerd[1889]: time="2026-04-23T23:19:12.649614565Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:12.650259 containerd[1889]: time="2026-04-23T23:19:12.650229780Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:138784d87c9c50f8e59412544da4cf4928d61ccbaf93b9f5898a3ba406871bfc\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"20392204\" in 1.793029936s" Apr 23 23:19:12.650291 containerd[1889]: time="2026-04-23T23:19:12.650262781Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:138784d87c9c50f8e59412544da4cf4928d61ccbaf93b9f5898a3ba406871bfc\"" Apr 23 23:19:12.651169 containerd[1889]: time="2026-04-23T23:19:12.651142014Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Apr 23 23:19:13.219240 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount553032024.mount: Deactivated successfully. Apr 23 23:19:13.243259 containerd[1889]: time="2026-04-23T23:19:13.243220878Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:13.246198 containerd[1889]: time="2026-04-23T23:19:13.246168299Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=268709" Apr 23 23:19:13.249621 containerd[1889]: time="2026-04-23T23:19:13.249302007Z" level=info msg="ImageCreate event name:\"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:13.252902 containerd[1889]: time="2026-04-23T23:19:13.252871084Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:13.253312 containerd[1889]: time="2026-04-23T23:19:13.253289540Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"267939\" in 602.116988ms" Apr 23 23:19:13.253389 containerd[1889]: time="2026-04-23T23:19:13.253375799Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\"" Apr 23 23:19:13.253841 containerd[1889]: time="2026-04-23T23:19:13.253816839Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\"" Apr 23 23:19:13.956385 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2385980931.mount: Deactivated successfully. Apr 23 23:19:14.377160 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Apr 23 23:19:14.383764 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:19:14.778755 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:19:14.786822 (kubelet)[2838]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:19:14.814627 kubelet[2838]: E0423 23:19:14.814570 2838 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:19:14.816342 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:19:14.816434 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:19:14.816690 systemd[1]: kubelet.service: Consumed 104ms CPU time, 106.9M memory peak. Apr 23 23:19:16.133951 containerd[1889]: time="2026-04-23T23:19:16.133901728Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.5-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:16.136801 containerd[1889]: time="2026-04-23T23:19:16.136607589Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.5-0: active requests=0, bytes read=21139658" Apr 23 23:19:16.139652 containerd[1889]: time="2026-04-23T23:19:16.139631341Z" level=info msg="ImageCreate event name:\"sha256:2c5f0dedd21c25ec3a6709934d22152d53ec50fe57b72d29e4450655e3d14d42\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:16.143934 containerd[1889]: time="2026-04-23T23:19:16.143906308Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:16.144581 containerd[1889]: time="2026-04-23T23:19:16.144554356Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.5-0\" with image id \"sha256:2c5f0dedd21c25ec3a6709934d22152d53ec50fe57b72d29e4450655e3d14d42\", repo tag \"registry.k8s.io/etcd:3.6.5-0\", repo digest \"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\", size \"21136588\" in 2.89071322s" Apr 23 23:19:16.144687 containerd[1889]: time="2026-04-23T23:19:16.144672417Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\" returns image reference \"sha256:2c5f0dedd21c25ec3a6709934d22152d53ec50fe57b72d29e4450655e3d14d42\"" Apr 23 23:19:19.238618 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:19:19.239019 systemd[1]: kubelet.service: Consumed 104ms CPU time, 106.9M memory peak. Apr 23 23:19:19.240717 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:19:19.260754 systemd[1]: Reload requested from client PID 2914 ('systemctl') (unit session-9.scope)... Apr 23 23:19:19.260768 systemd[1]: Reloading... Apr 23 23:19:19.358616 zram_generator::config[2982]: No configuration found. Apr 23 23:19:19.489299 systemd[1]: Reloading finished in 228 ms. Apr 23 23:19:19.528042 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 23 23:19:19.528098 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 23 23:19:19.528320 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:19:19.528361 systemd[1]: kubelet.service: Consumed 71ms CPU time, 94.9M memory peak. Apr 23 23:19:19.529443 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:19:19.767418 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:19:19.777803 (kubelet)[3028]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 23 23:19:19.942881 kubelet[3028]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 23 23:19:19.943126 kubelet[3028]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 23 23:19:19.943747 kubelet[3028]: I0423 23:19:19.943715 3028 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 23 23:19:20.359731 kubelet[3028]: I0423 23:19:20.359699 3028 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Apr 23 23:19:20.359878 kubelet[3028]: I0423 23:19:20.359868 3028 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 23 23:19:20.359940 kubelet[3028]: I0423 23:19:20.359932 3028 watchdog_linux.go:95] "Systemd watchdog is not enabled" Apr 23 23:19:20.359988 kubelet[3028]: I0423 23:19:20.359977 3028 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 23 23:19:20.360196 kubelet[3028]: I0423 23:19:20.360180 3028 server.go:956] "Client rotation is on, will bootstrap in background" Apr 23 23:19:20.370604 kubelet[3028]: E0423 23:19:20.370568 3028 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.13:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 23 23:19:20.372251 kubelet[3028]: I0423 23:19:20.372228 3028 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 23 23:19:20.375610 kubelet[3028]: I0423 23:19:20.375576 3028 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 23 23:19:20.377907 kubelet[3028]: I0423 23:19:20.377890 3028 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Apr 23 23:19:20.378060 kubelet[3028]: I0423 23:19:20.378032 3028 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 23 23:19:20.378167 kubelet[3028]: I0423 23:19:20.378058 3028 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459.2.4-n-8b75ae451f","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 23 23:19:20.378167 kubelet[3028]: I0423 23:19:20.378166 3028 topology_manager.go:138] "Creating topology manager with none policy" Apr 23 23:19:20.378249 kubelet[3028]: I0423 23:19:20.378173 3028 container_manager_linux.go:306] "Creating device plugin manager" Apr 23 23:19:20.378249 kubelet[3028]: I0423 23:19:20.378240 3028 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Apr 23 23:19:20.383426 kubelet[3028]: I0423 23:19:20.383409 3028 state_mem.go:36] "Initialized new in-memory state store" Apr 23 23:19:20.384481 kubelet[3028]: I0423 23:19:20.384464 3028 kubelet.go:475] "Attempting to sync node with API server" Apr 23 23:19:20.384481 kubelet[3028]: I0423 23:19:20.384483 3028 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 23 23:19:20.384847 kubelet[3028]: E0423 23:19:20.384823 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.13:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4459.2.4-n-8b75ae451f&limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 23 23:19:20.385397 kubelet[3028]: I0423 23:19:20.385380 3028 kubelet.go:387] "Adding apiserver pod source" Apr 23 23:19:20.385424 kubelet[3028]: I0423 23:19:20.385401 3028 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 23 23:19:20.385960 kubelet[3028]: E0423 23:19:20.385937 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.13:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 23 23:19:20.387447 kubelet[3028]: I0423 23:19:20.387423 3028 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Apr 23 23:19:20.388158 kubelet[3028]: I0423 23:19:20.387786 3028 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 23 23:19:20.388158 kubelet[3028]: I0423 23:19:20.387806 3028 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Apr 23 23:19:20.388158 kubelet[3028]: W0423 23:19:20.387838 3028 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 23 23:19:20.389502 kubelet[3028]: I0423 23:19:20.389472 3028 server.go:1262] "Started kubelet" Apr 23 23:19:20.390091 kubelet[3028]: I0423 23:19:20.390071 3028 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 23 23:19:20.390834 kubelet[3028]: I0423 23:19:20.390817 3028 server.go:310] "Adding debug handlers to kubelet server" Apr 23 23:19:20.392493 kubelet[3028]: I0423 23:19:20.392451 3028 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 23 23:19:20.392552 kubelet[3028]: I0423 23:19:20.392504 3028 server_v1.go:49] "podresources" method="list" useActivePods=true Apr 23 23:19:20.393313 kubelet[3028]: I0423 23:19:20.392767 3028 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 23 23:19:20.393715 kubelet[3028]: E0423 23:19:20.392865 3028 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.13:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.13:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4459.2.4-n-8b75ae451f.18a91fae0dba42b4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4459.2.4-n-8b75ae451f,UID:ci-4459.2.4-n-8b75ae451f,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4459.2.4-n-8b75ae451f,},FirstTimestamp:2026-04-23 23:19:20.389452468 +0000 UTC m=+0.609218747,LastTimestamp:2026-04-23 23:19:20.389452468 +0000 UTC m=+0.609218747,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459.2.4-n-8b75ae451f,}" Apr 23 23:19:20.394796 kubelet[3028]: I0423 23:19:20.394640 3028 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 23 23:19:20.394796 kubelet[3028]: I0423 23:19:20.394780 3028 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 23 23:19:20.395862 kubelet[3028]: I0423 23:19:20.395839 3028 volume_manager.go:313] "Starting Kubelet Volume Manager" Apr 23 23:19:20.395924 kubelet[3028]: I0423 23:19:20.395899 3028 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 23 23:19:20.395949 kubelet[3028]: I0423 23:19:20.395934 3028 reconciler.go:29] "Reconciler: start to sync state" Apr 23 23:19:20.396919 kubelet[3028]: E0423 23:19:20.396155 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.13:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 23 23:19:20.396919 kubelet[3028]: E0423 23:19:20.396288 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:20.396919 kubelet[3028]: E0423 23:19:20.396345 3028 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.4-n-8b75ae451f?timeout=10s\": dial tcp 10.0.0.13:6443: connect: connection refused" interval="200ms" Apr 23 23:19:20.397366 kubelet[3028]: I0423 23:19:20.397346 3028 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 23 23:19:20.399094 kubelet[3028]: E0423 23:19:20.399072 3028 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 23 23:19:20.399903 kubelet[3028]: I0423 23:19:20.399532 3028 factory.go:223] Registration of the containerd container factory successfully Apr 23 23:19:20.400069 kubelet[3028]: I0423 23:19:20.400058 3028 factory.go:223] Registration of the systemd container factory successfully Apr 23 23:19:20.425310 kubelet[3028]: I0423 23:19:20.425299 3028 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 23 23:19:20.425844 kubelet[3028]: I0423 23:19:20.425682 3028 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 23 23:19:20.425844 kubelet[3028]: I0423 23:19:20.425702 3028 state_mem.go:36] "Initialized new in-memory state store" Apr 23 23:19:20.434120 kubelet[3028]: I0423 23:19:20.434105 3028 policy_none.go:49] "None policy: Start" Apr 23 23:19:20.434204 kubelet[3028]: I0423 23:19:20.434196 3028 memory_manager.go:187] "Starting memorymanager" policy="None" Apr 23 23:19:20.434253 kubelet[3028]: I0423 23:19:20.434245 3028 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Apr 23 23:19:20.441628 kubelet[3028]: I0423 23:19:20.440405 3028 policy_none.go:47] "Start" Apr 23 23:19:20.449063 kubelet[3028]: I0423 23:19:20.448814 3028 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Apr 23 23:19:20.449406 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 23 23:19:20.451605 kubelet[3028]: I0423 23:19:20.451534 3028 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Apr 23 23:19:20.451605 kubelet[3028]: I0423 23:19:20.451553 3028 status_manager.go:244] "Starting to sync pod status with apiserver" Apr 23 23:19:20.451605 kubelet[3028]: I0423 23:19:20.451587 3028 kubelet.go:2428] "Starting kubelet main sync loop" Apr 23 23:19:20.451701 kubelet[3028]: E0423 23:19:20.451686 3028 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 23 23:19:20.452290 kubelet[3028]: E0423 23:19:20.452147 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.13:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 23 23:19:20.455891 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 23 23:19:20.459763 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 23 23:19:20.470145 kubelet[3028]: E0423 23:19:20.470121 3028 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 23 23:19:20.470296 kubelet[3028]: I0423 23:19:20.470277 3028 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 23 23:19:20.470335 kubelet[3028]: I0423 23:19:20.470290 3028 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 23 23:19:20.471088 kubelet[3028]: I0423 23:19:20.471062 3028 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 23 23:19:20.472249 kubelet[3028]: E0423 23:19:20.472155 3028 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 23 23:19:20.472249 kubelet[3028]: E0423 23:19:20.472185 3028 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:20.572104 kubelet[3028]: I0423 23:19:20.571840 3028 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.572194 kubelet[3028]: E0423 23:19:20.572154 3028 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.13:6443/api/v1/nodes\": dial tcp 10.0.0.13:6443: connect: connection refused" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.596440 kubelet[3028]: I0423 23:19:20.596412 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-ca-certs\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.596797 kubelet[3028]: E0423 23:19:20.596765 3028 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.4-n-8b75ae451f?timeout=10s\": dial tcp 10.0.0.13:6443: connect: connection refused" interval="400ms" Apr 23 23:19:20.697131 kubelet[3028]: I0423 23:19:20.696463 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-flexvolume-dir\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.697131 kubelet[3028]: I0423 23:19:20.696494 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-k8s-certs\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.697131 kubelet[3028]: I0423 23:19:20.696505 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-kubeconfig\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.697131 kubelet[3028]: I0423 23:19:20.696529 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.701655 systemd[1]: Created slice kubepods-burstable-podbdfc4518aba338a6ec858db5f6bdd549.slice - libcontainer container kubepods-burstable-podbdfc4518aba338a6ec858db5f6bdd549.slice. Apr 23 23:19:20.708142 kubelet[3028]: E0423 23:19:20.708120 3028 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.715099 systemd[1]: Created slice kubepods-burstable-poddd2e45e9288d406fdd1bb6209ffe6ed6.slice - libcontainer container kubepods-burstable-poddd2e45e9288d406fdd1bb6209ffe6ed6.slice. Apr 23 23:19:20.717010 kubelet[3028]: E0423 23:19:20.716990 3028 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.741949 systemd[1]: Created slice kubepods-burstable-podb6258660ea9771448c3d31fed4c6e202.slice - libcontainer container kubepods-burstable-podb6258660ea9771448c3d31fed4c6e202.slice. Apr 23 23:19:20.743479 kubelet[3028]: E0423 23:19:20.743453 3028 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.773684 kubelet[3028]: I0423 23:19:20.773667 3028 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.773929 kubelet[3028]: E0423 23:19:20.773903 3028 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.13:6443/api/v1/nodes\": dial tcp 10.0.0.13:6443: connect: connection refused" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.797287 kubelet[3028]: I0423 23:19:20.797209 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b6258660ea9771448c3d31fed4c6e202-k8s-certs\") pod \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" (UID: \"b6258660ea9771448c3d31fed4c6e202\") " pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.797329 kubelet[3028]: I0423 23:19:20.797301 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b6258660ea9771448c3d31fed4c6e202-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" (UID: \"b6258660ea9771448c3d31fed4c6e202\") " pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.797329 kubelet[3028]: I0423 23:19:20.797326 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b6258660ea9771448c3d31fed4c6e202-ca-certs\") pod \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" (UID: \"b6258660ea9771448c3d31fed4c6e202\") " pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.797519 kubelet[3028]: I0423 23:19:20.797348 3028 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dd2e45e9288d406fdd1bb6209ffe6ed6-kubeconfig\") pod \"kube-scheduler-ci-4459.2.4-n-8b75ae451f\" (UID: \"dd2e45e9288d406fdd1bb6209ffe6ed6\") " pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:20.997540 kubelet[3028]: E0423 23:19:20.997444 3028 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.4-n-8b75ae451f?timeout=10s\": dial tcp 10.0.0.13:6443: connect: connection refused" interval="800ms" Apr 23 23:19:21.176080 kubelet[3028]: I0423 23:19:21.176022 3028 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:21.176585 kubelet[3028]: E0423 23:19:21.176553 3028 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.13:6443/api/v1/nodes\": dial tcp 10.0.0.13:6443: connect: connection refused" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:21.737901 containerd[1889]: time="2026-04-23T23:19:21.737806645Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459.2.4-n-8b75ae451f,Uid:bdfc4518aba338a6ec858db5f6bdd549,Namespace:kube-system,Attempt:0,}" Apr 23 23:19:21.742528 containerd[1889]: time="2026-04-23T23:19:21.742492676Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459.2.4-n-8b75ae451f,Uid:dd2e45e9288d406fdd1bb6209ffe6ed6,Namespace:kube-system,Attempt:0,}" Apr 23 23:19:21.747749 containerd[1889]: time="2026-04-23T23:19:21.747723151Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459.2.4-n-8b75ae451f,Uid:b6258660ea9771448c3d31fed4c6e202,Namespace:kube-system,Attempt:0,}" Apr 23 23:19:21.765377 kubelet[3028]: E0423 23:19:21.765347 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.13:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4459.2.4-n-8b75ae451f&limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 23 23:19:21.798888 kubelet[3028]: E0423 23:19:21.798857 3028 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.4-n-8b75ae451f?timeout=10s\": dial tcp 10.0.0.13:6443: connect: connection refused" interval="1.6s" Apr 23 23:19:21.950847 kubelet[3028]: E0423 23:19:21.950811 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.13:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 23 23:19:21.967244 kubelet[3028]: E0423 23:19:21.967219 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.13:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 23 23:19:21.978172 kubelet[3028]: I0423 23:19:21.978151 3028 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:21.978378 kubelet[3028]: E0423 23:19:21.978355 3028 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.13:6443/api/v1/nodes\": dial tcp 10.0.0.13:6443: connect: connection refused" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:22.041186 kubelet[3028]: E0423 23:19:22.041113 3028 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.13:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 23 23:19:22.370755 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1736903457.mount: Deactivated successfully. Apr 23 23:19:22.393284 containerd[1889]: time="2026-04-23T23:19:22.393244382Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:19:22.404989 containerd[1889]: time="2026-04-23T23:19:22.404954275Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Apr 23 23:19:22.407776 containerd[1889]: time="2026-04-23T23:19:22.407743483Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:19:22.411256 containerd[1889]: time="2026-04-23T23:19:22.410815302Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:19:22.415991 containerd[1889]: time="2026-04-23T23:19:22.415970814Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Apr 23 23:19:22.419160 containerd[1889]: time="2026-04-23T23:19:22.419127268Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:19:22.422683 containerd[1889]: time="2026-04-23T23:19:22.422658408Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:19:22.423222 containerd[1889]: time="2026-04-23T23:19:22.423200420Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 681.179346ms" Apr 23 23:19:22.425307 containerd[1889]: time="2026-04-23T23:19:22.425283730Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Apr 23 23:19:22.428740 containerd[1889]: time="2026-04-23T23:19:22.428711154Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 671.53157ms" Apr 23 23:19:22.452861 containerd[1889]: time="2026-04-23T23:19:22.452824975Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 702.031949ms" Apr 23 23:19:22.457142 kubelet[3028]: E0423 23:19:22.457103 3028 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.13:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.13:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 23 23:19:22.473287 containerd[1889]: time="2026-04-23T23:19:22.473240185Z" level=info msg="connecting to shim 67af67b885d14a22150453ee8c9d6f27a038de8b0ed491ac57cacde1250f1a7c" address="unix:///run/containerd/s/810a0ac17a21a481a79deeefc2bee327eb36deed4b65cbf275d04bd5f6051952" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:22.480133 containerd[1889]: time="2026-04-23T23:19:22.480087961Z" level=info msg="connecting to shim 37d46060a30328ff134cc683da44b4ef01445db3015edf451a2c33df52dd0f33" address="unix:///run/containerd/s/c60ae3af113da414dace505c6e50278cd3ecc0d232ac2f55ab2b12a7abb6f146" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:22.495729 systemd[1]: Started cri-containerd-67af67b885d14a22150453ee8c9d6f27a038de8b0ed491ac57cacde1250f1a7c.scope - libcontainer container 67af67b885d14a22150453ee8c9d6f27a038de8b0ed491ac57cacde1250f1a7c. Apr 23 23:19:22.501433 systemd[1]: Started cri-containerd-37d46060a30328ff134cc683da44b4ef01445db3015edf451a2c33df52dd0f33.scope - libcontainer container 37d46060a30328ff134cc683da44b4ef01445db3015edf451a2c33df52dd0f33. Apr 23 23:19:22.503616 containerd[1889]: time="2026-04-23T23:19:22.503568574Z" level=info msg="connecting to shim 52ece136a70dd1aa4376b2b9e55dc7b26fcb21586a37dfe6f00cd8d4ce9c41f5" address="unix:///run/containerd/s/c096da39671d92351127bfd759294556bedfd907d27c475cad789dc248470a3b" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:22.531729 systemd[1]: Started cri-containerd-52ece136a70dd1aa4376b2b9e55dc7b26fcb21586a37dfe6f00cd8d4ce9c41f5.scope - libcontainer container 52ece136a70dd1aa4376b2b9e55dc7b26fcb21586a37dfe6f00cd8d4ce9c41f5. Apr 23 23:19:22.543568 containerd[1889]: time="2026-04-23T23:19:22.543436295Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459.2.4-n-8b75ae451f,Uid:bdfc4518aba338a6ec858db5f6bdd549,Namespace:kube-system,Attempt:0,} returns sandbox id \"67af67b885d14a22150453ee8c9d6f27a038de8b0ed491ac57cacde1250f1a7c\"" Apr 23 23:19:22.555091 containerd[1889]: time="2026-04-23T23:19:22.554503677Z" level=info msg="CreateContainer within sandbox \"67af67b885d14a22150453ee8c9d6f27a038de8b0ed491ac57cacde1250f1a7c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 23 23:19:22.555679 containerd[1889]: time="2026-04-23T23:19:22.555656112Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459.2.4-n-8b75ae451f,Uid:b6258660ea9771448c3d31fed4c6e202,Namespace:kube-system,Attempt:0,} returns sandbox id \"37d46060a30328ff134cc683da44b4ef01445db3015edf451a2c33df52dd0f33\"" Apr 23 23:19:22.564788 containerd[1889]: time="2026-04-23T23:19:22.564746611Z" level=info msg="CreateContainer within sandbox \"37d46060a30328ff134cc683da44b4ef01445db3015edf451a2c33df52dd0f33\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 23 23:19:22.576580 containerd[1889]: time="2026-04-23T23:19:22.576543516Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459.2.4-n-8b75ae451f,Uid:dd2e45e9288d406fdd1bb6209ffe6ed6,Namespace:kube-system,Attempt:0,} returns sandbox id \"52ece136a70dd1aa4376b2b9e55dc7b26fcb21586a37dfe6f00cd8d4ce9c41f5\"" Apr 23 23:19:22.583538 containerd[1889]: time="2026-04-23T23:19:22.583463822Z" level=info msg="Container 32a56ccf60a7f8f2619ea48cd96a64a8dce4faeb2f99d466f929ea7ac6d6a694: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:22.585140 containerd[1889]: time="2026-04-23T23:19:22.584688444Z" level=info msg="CreateContainer within sandbox \"52ece136a70dd1aa4376b2b9e55dc7b26fcb21586a37dfe6f00cd8d4ce9c41f5\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 23 23:19:22.598196 containerd[1889]: time="2026-04-23T23:19:22.598162643Z" level=info msg="Container e6dc74a2f3441756ca28dd92ecb23c93eceb220eeec899df690049d2302bfee0: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:22.662662 containerd[1889]: time="2026-04-23T23:19:22.662631251Z" level=info msg="Container df83b459df0afb95e5da694beeeded11d39f81f6bee046ad1a54b56f876fef5d: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:22.675501 containerd[1889]: time="2026-04-23T23:19:22.675471563Z" level=info msg="CreateContainer within sandbox \"67af67b885d14a22150453ee8c9d6f27a038de8b0ed491ac57cacde1250f1a7c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"32a56ccf60a7f8f2619ea48cd96a64a8dce4faeb2f99d466f929ea7ac6d6a694\"" Apr 23 23:19:22.675986 containerd[1889]: time="2026-04-23T23:19:22.675958261Z" level=info msg="CreateContainer within sandbox \"37d46060a30328ff134cc683da44b4ef01445db3015edf451a2c33df52dd0f33\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"e6dc74a2f3441756ca28dd92ecb23c93eceb220eeec899df690049d2302bfee0\"" Apr 23 23:19:22.675986 containerd[1889]: time="2026-04-23T23:19:22.676095986Z" level=info msg="StartContainer for \"32a56ccf60a7f8f2619ea48cd96a64a8dce4faeb2f99d466f929ea7ac6d6a694\"" Apr 23 23:19:22.676854 containerd[1889]: time="2026-04-23T23:19:22.676828109Z" level=info msg="connecting to shim 32a56ccf60a7f8f2619ea48cd96a64a8dce4faeb2f99d466f929ea7ac6d6a694" address="unix:///run/containerd/s/810a0ac17a21a481a79deeefc2bee327eb36deed4b65cbf275d04bd5f6051952" protocol=ttrpc version=3 Apr 23 23:19:22.677746 containerd[1889]: time="2026-04-23T23:19:22.677280966Z" level=info msg="StartContainer for \"e6dc74a2f3441756ca28dd92ecb23c93eceb220eeec899df690049d2302bfee0\"" Apr 23 23:19:22.678099 containerd[1889]: time="2026-04-23T23:19:22.678069332Z" level=info msg="connecting to shim e6dc74a2f3441756ca28dd92ecb23c93eceb220eeec899df690049d2302bfee0" address="unix:///run/containerd/s/c60ae3af113da414dace505c6e50278cd3ecc0d232ac2f55ab2b12a7abb6f146" protocol=ttrpc version=3 Apr 23 23:19:22.686481 containerd[1889]: time="2026-04-23T23:19:22.686450685Z" level=info msg="CreateContainer within sandbox \"52ece136a70dd1aa4376b2b9e55dc7b26fcb21586a37dfe6f00cd8d4ce9c41f5\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"df83b459df0afb95e5da694beeeded11d39f81f6bee046ad1a54b56f876fef5d\"" Apr 23 23:19:22.687028 containerd[1889]: time="2026-04-23T23:19:22.687005554Z" level=info msg="StartContainer for \"df83b459df0afb95e5da694beeeded11d39f81f6bee046ad1a54b56f876fef5d\"" Apr 23 23:19:22.687694 containerd[1889]: time="2026-04-23T23:19:22.687668650Z" level=info msg="connecting to shim df83b459df0afb95e5da694beeeded11d39f81f6bee046ad1a54b56f876fef5d" address="unix:///run/containerd/s/c096da39671d92351127bfd759294556bedfd907d27c475cad789dc248470a3b" protocol=ttrpc version=3 Apr 23 23:19:22.695719 systemd[1]: Started cri-containerd-32a56ccf60a7f8f2619ea48cd96a64a8dce4faeb2f99d466f929ea7ac6d6a694.scope - libcontainer container 32a56ccf60a7f8f2619ea48cd96a64a8dce4faeb2f99d466f929ea7ac6d6a694. Apr 23 23:19:22.696381 systemd[1]: Started cri-containerd-e6dc74a2f3441756ca28dd92ecb23c93eceb220eeec899df690049d2302bfee0.scope - libcontainer container e6dc74a2f3441756ca28dd92ecb23c93eceb220eeec899df690049d2302bfee0. Apr 23 23:19:22.711736 systemd[1]: Started cri-containerd-df83b459df0afb95e5da694beeeded11d39f81f6bee046ad1a54b56f876fef5d.scope - libcontainer container df83b459df0afb95e5da694beeeded11d39f81f6bee046ad1a54b56f876fef5d. Apr 23 23:19:22.755030 containerd[1889]: time="2026-04-23T23:19:22.753944926Z" level=info msg="StartContainer for \"e6dc74a2f3441756ca28dd92ecb23c93eceb220eeec899df690049d2302bfee0\" returns successfully" Apr 23 23:19:22.765115 containerd[1889]: time="2026-04-23T23:19:22.765079334Z" level=info msg="StartContainer for \"32a56ccf60a7f8f2619ea48cd96a64a8dce4faeb2f99d466f929ea7ac6d6a694\" returns successfully" Apr 23 23:19:22.781107 containerd[1889]: time="2026-04-23T23:19:22.780645291Z" level=info msg="StartContainer for \"df83b459df0afb95e5da694beeeded11d39f81f6bee046ad1a54b56f876fef5d\" returns successfully" Apr 23 23:19:23.462011 kubelet[3028]: E0423 23:19:23.461827 3028 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:23.465657 kubelet[3028]: E0423 23:19:23.465558 3028 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:23.466756 kubelet[3028]: E0423 23:19:23.466646 3028 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:23.580047 kubelet[3028]: I0423 23:19:23.580025 3028 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:23.650413 kubelet[3028]: E0423 23:19:23.650380 3028 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4459.2.4-n-8b75ae451f\" not found" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:23.726747 kubelet[3028]: I0423 23:19:23.726255 3028 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:23.726747 kubelet[3028]: E0423 23:19:23.726369 3028 kubelet_node_status.go:486] "Error updating node status, will retry" err="error getting node \"ci-4459.2.4-n-8b75ae451f\": node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:23.745027 kubelet[3028]: E0423 23:19:23.744843 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:23.845942 kubelet[3028]: E0423 23:19:23.845916 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:23.946761 kubelet[3028]: E0423 23:19:23.946731 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:24.046878 kubelet[3028]: E0423 23:19:24.046781 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:24.147801 kubelet[3028]: E0423 23:19:24.147771 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:24.248206 kubelet[3028]: E0423 23:19:24.248179 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:24.348524 kubelet[3028]: E0423 23:19:24.348440 3028 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4459.2.4-n-8b75ae451f\" not found" Apr 23 23:19:24.467379 kubelet[3028]: I0423 23:19:24.467341 3028 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.467892 kubelet[3028]: I0423 23:19:24.467779 3028 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.471911 kubelet[3028]: E0423 23:19:24.471881 3028 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459.2.4-n-8b75ae451f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.472197 kubelet[3028]: E0423 23:19:24.472172 3028 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.496532 kubelet[3028]: I0423 23:19:24.496508 3028 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.498022 kubelet[3028]: E0423 23:19:24.497997 3028 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.498022 kubelet[3028]: I0423 23:19:24.498017 3028 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.499222 kubelet[3028]: E0423 23:19:24.499197 3028 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.499222 kubelet[3028]: I0423 23:19:24.499217 3028 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:24.500330 kubelet[3028]: E0423 23:19:24.500307 3028 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459.2.4-n-8b75ae451f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:25.388301 kubelet[3028]: I0423 23:19:25.388273 3028 apiserver.go:52] "Watching apiserver" Apr 23 23:19:25.396171 kubelet[3028]: I0423 23:19:25.396145 3028 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 23 23:19:25.864679 systemd[1]: Reload requested from client PID 3316 ('systemctl') (unit session-9.scope)... Apr 23 23:19:25.864696 systemd[1]: Reloading... Apr 23 23:19:25.936652 zram_generator::config[3366]: No configuration found. Apr 23 23:19:26.097811 systemd[1]: Reloading finished in 232 ms. Apr 23 23:19:26.121671 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:19:26.132899 systemd[1]: kubelet.service: Deactivated successfully. Apr 23 23:19:26.133063 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:19:26.133100 systemd[1]: kubelet.service: Consumed 686ms CPU time, 121.5M memory peak. Apr 23 23:19:26.135090 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:19:26.236680 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:19:26.244821 (kubelet)[3427]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 23 23:19:26.271162 kubelet[3427]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 23 23:19:26.271162 kubelet[3427]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 23 23:19:26.271162 kubelet[3427]: I0423 23:19:26.271067 3427 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 23 23:19:26.276707 kubelet[3427]: I0423 23:19:26.276688 3427 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Apr 23 23:19:26.276815 kubelet[3427]: I0423 23:19:26.276795 3427 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 23 23:19:26.276968 kubelet[3427]: I0423 23:19:26.276873 3427 watchdog_linux.go:95] "Systemd watchdog is not enabled" Apr 23 23:19:26.276968 kubelet[3427]: I0423 23:19:26.276894 3427 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 23 23:19:26.277190 kubelet[3427]: I0423 23:19:26.277176 3427 server.go:956] "Client rotation is on, will bootstrap in background" Apr 23 23:19:26.278314 kubelet[3427]: I0423 23:19:26.278298 3427 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 23 23:19:26.281155 kubelet[3427]: I0423 23:19:26.281120 3427 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 23 23:19:26.283273 kubelet[3427]: I0423 23:19:26.283203 3427 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 23 23:19:26.285519 kubelet[3427]: I0423 23:19:26.285505 3427 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Apr 23 23:19:26.285790 kubelet[3427]: I0423 23:19:26.285772 3427 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 23 23:19:26.286311 kubelet[3427]: I0423 23:19:26.285855 3427 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459.2.4-n-8b75ae451f","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 23 23:19:26.286429 kubelet[3427]: I0423 23:19:26.286419 3427 topology_manager.go:138] "Creating topology manager with none policy" Apr 23 23:19:26.286478 kubelet[3427]: I0423 23:19:26.286469 3427 container_manager_linux.go:306] "Creating device plugin manager" Apr 23 23:19:26.286533 kubelet[3427]: I0423 23:19:26.286525 3427 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Apr 23 23:19:26.286733 kubelet[3427]: I0423 23:19:26.286719 3427 state_mem.go:36] "Initialized new in-memory state store" Apr 23 23:19:26.286919 kubelet[3427]: I0423 23:19:26.286909 3427 kubelet.go:475] "Attempting to sync node with API server" Apr 23 23:19:26.286999 kubelet[3427]: I0423 23:19:26.286991 3427 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 23 23:19:26.287058 kubelet[3427]: I0423 23:19:26.287052 3427 kubelet.go:387] "Adding apiserver pod source" Apr 23 23:19:26.287223 kubelet[3427]: I0423 23:19:26.287118 3427 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 23 23:19:26.294007 kubelet[3427]: I0423 23:19:26.293991 3427 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Apr 23 23:19:26.294522 kubelet[3427]: I0423 23:19:26.294507 3427 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 23 23:19:26.294624 kubelet[3427]: I0423 23:19:26.294609 3427 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Apr 23 23:19:26.297993 kubelet[3427]: I0423 23:19:26.297263 3427 server.go:1262] "Started kubelet" Apr 23 23:19:26.300120 kubelet[3427]: I0423 23:19:26.300103 3427 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 23 23:19:26.300904 kubelet[3427]: I0423 23:19:26.300730 3427 server.go:310] "Adding debug handlers to kubelet server" Apr 23 23:19:26.301273 kubelet[3427]: I0423 23:19:26.301228 3427 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 23 23:19:26.301332 kubelet[3427]: I0423 23:19:26.301280 3427 server_v1.go:49] "podresources" method="list" useActivePods=true Apr 23 23:19:26.301842 kubelet[3427]: I0423 23:19:26.301787 3427 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 23 23:19:26.303692 kubelet[3427]: I0423 23:19:26.303666 3427 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 23 23:19:26.305186 kubelet[3427]: I0423 23:19:26.305123 3427 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 23 23:19:26.306751 kubelet[3427]: E0423 23:19:26.306474 3427 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 23 23:19:26.308503 kubelet[3427]: I0423 23:19:26.308481 3427 volume_manager.go:313] "Starting Kubelet Volume Manager" Apr 23 23:19:26.309468 kubelet[3427]: I0423 23:19:26.308928 3427 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 23 23:19:26.309468 kubelet[3427]: I0423 23:19:26.309039 3427 reconciler.go:29] "Reconciler: start to sync state" Apr 23 23:19:26.310841 kubelet[3427]: I0423 23:19:26.310808 3427 factory.go:223] Registration of the systemd container factory successfully Apr 23 23:19:26.310901 kubelet[3427]: I0423 23:19:26.310879 3427 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 23 23:19:26.313164 kubelet[3427]: I0423 23:19:26.313137 3427 factory.go:223] Registration of the containerd container factory successfully Apr 23 23:19:26.315292 kubelet[3427]: I0423 23:19:26.314984 3427 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Apr 23 23:19:26.315798 kubelet[3427]: I0423 23:19:26.315774 3427 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Apr 23 23:19:26.315798 kubelet[3427]: I0423 23:19:26.315793 3427 status_manager.go:244] "Starting to sync pod status with apiserver" Apr 23 23:19:26.315881 kubelet[3427]: I0423 23:19:26.315808 3427 kubelet.go:2428] "Starting kubelet main sync loop" Apr 23 23:19:26.315881 kubelet[3427]: E0423 23:19:26.315841 3427 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 23 23:19:26.352313 kubelet[3427]: I0423 23:19:26.352291 3427 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 23 23:19:26.352313 kubelet[3427]: I0423 23:19:26.352305 3427 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 23 23:19:26.352398 kubelet[3427]: I0423 23:19:26.352322 3427 state_mem.go:36] "Initialized new in-memory state store" Apr 23 23:19:26.352415 kubelet[3427]: I0423 23:19:26.352406 3427 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 23 23:19:26.352431 kubelet[3427]: I0423 23:19:26.352413 3427 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 23 23:19:26.352431 kubelet[3427]: I0423 23:19:26.352423 3427 policy_none.go:49] "None policy: Start" Apr 23 23:19:26.352431 kubelet[3427]: I0423 23:19:26.352429 3427 memory_manager.go:187] "Starting memorymanager" policy="None" Apr 23 23:19:26.352478 kubelet[3427]: I0423 23:19:26.352436 3427 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Apr 23 23:19:26.352493 kubelet[3427]: I0423 23:19:26.352490 3427 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Apr 23 23:19:26.352506 kubelet[3427]: I0423 23:19:26.352495 3427 policy_none.go:47] "Start" Apr 23 23:19:26.356519 kubelet[3427]: E0423 23:19:26.356504 3427 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 23 23:19:26.356916 kubelet[3427]: I0423 23:19:26.356903 3427 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 23 23:19:26.357014 kubelet[3427]: I0423 23:19:26.356991 3427 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 23 23:19:26.357923 kubelet[3427]: I0423 23:19:26.357897 3427 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 23 23:19:26.360645 kubelet[3427]: E0423 23:19:26.360627 3427 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 23 23:19:26.416899 kubelet[3427]: I0423 23:19:26.416665 3427 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.416899 kubelet[3427]: I0423 23:19:26.416700 3427 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.416899 kubelet[3427]: I0423 23:19:26.416825 3427 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.424032 kubelet[3427]: I0423 23:19:26.423698 3427 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 23 23:19:26.428208 kubelet[3427]: I0423 23:19:26.428193 3427 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 23 23:19:26.428657 kubelet[3427]: I0423 23:19:26.428582 3427 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 23 23:19:26.462508 kubelet[3427]: I0423 23:19:26.462429 3427 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.482779 kubelet[3427]: I0423 23:19:26.482752 3427 kubelet_node_status.go:124] "Node was previously registered" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.482855 kubelet[3427]: I0423 23:19:26.482815 3427 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609615 kubelet[3427]: I0423 23:19:26.609427 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609615 kubelet[3427]: I0423 23:19:26.609464 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b6258660ea9771448c3d31fed4c6e202-ca-certs\") pod \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" (UID: \"b6258660ea9771448c3d31fed4c6e202\") " pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609615 kubelet[3427]: I0423 23:19:26.609478 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-ca-certs\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609615 kubelet[3427]: I0423 23:19:26.609492 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-k8s-certs\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609615 kubelet[3427]: I0423 23:19:26.609507 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dd2e45e9288d406fdd1bb6209ffe6ed6-kubeconfig\") pod \"kube-scheduler-ci-4459.2.4-n-8b75ae451f\" (UID: \"dd2e45e9288d406fdd1bb6209ffe6ed6\") " pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609771 kubelet[3427]: I0423 23:19:26.609515 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b6258660ea9771448c3d31fed4c6e202-k8s-certs\") pod \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" (UID: \"b6258660ea9771448c3d31fed4c6e202\") " pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609771 kubelet[3427]: I0423 23:19:26.609523 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b6258660ea9771448c3d31fed4c6e202-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" (UID: \"b6258660ea9771448c3d31fed4c6e202\") " pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609771 kubelet[3427]: I0423 23:19:26.609532 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-flexvolume-dir\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:26.609771 kubelet[3427]: I0423 23:19:26.609543 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bdfc4518aba338a6ec858db5f6bdd549-kubeconfig\") pod \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" (UID: \"bdfc4518aba338a6ec858db5f6bdd549\") " pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:27.288271 kubelet[3427]: I0423 23:19:27.288238 3427 apiserver.go:52] "Watching apiserver" Apr 23 23:19:27.309455 kubelet[3427]: I0423 23:19:27.309417 3427 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 23 23:19:27.337478 kubelet[3427]: I0423 23:19:27.337454 3427 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:27.338943 kubelet[3427]: I0423 23:19:27.338920 3427 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:27.339217 kubelet[3427]: I0423 23:19:27.339195 3427 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:27.355436 kubelet[3427]: I0423 23:19:27.355254 3427 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 23 23:19:27.356567 kubelet[3427]: E0423 23:19:27.356540 3427 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459.2.4-n-8b75ae451f\" already exists" pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:27.361398 kubelet[3427]: I0423 23:19:27.361366 3427 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 23 23:19:27.361657 kubelet[3427]: E0423 23:19:27.361635 3427 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.4-n-8b75ae451f\" already exists" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:27.361978 kubelet[3427]: I0423 23:19:27.361743 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" podStartSLOduration=1.361735055 podStartE2EDuration="1.361735055s" podCreationTimestamp="2026-04-23 23:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:19:27.343691179 +0000 UTC m=+1.095813210" watchObservedRunningTime="2026-04-23 23:19:27.361735055 +0000 UTC m=+1.113857086" Apr 23 23:19:27.362547 kubelet[3427]: I0423 23:19:27.362514 3427 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 23 23:19:27.362624 kubelet[3427]: E0423 23:19:27.362551 3427 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459.2.4-n-8b75ae451f\" already exists" pod="kube-system/kube-scheduler-ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:27.374618 kubelet[3427]: I0423 23:19:27.373692 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4459.2.4-n-8b75ae451f" podStartSLOduration=1.373684173 podStartE2EDuration="1.373684173s" podCreationTimestamp="2026-04-23 23:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:19:27.373567992 +0000 UTC m=+1.125690023" watchObservedRunningTime="2026-04-23 23:19:27.373684173 +0000 UTC m=+1.125806204" Apr 23 23:19:27.374618 kubelet[3427]: I0423 23:19:27.373753 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4459.2.4-n-8b75ae451f" podStartSLOduration=1.373749991 podStartE2EDuration="1.373749991s" podCreationTimestamp="2026-04-23 23:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:19:27.362656443 +0000 UTC m=+1.114778482" watchObservedRunningTime="2026-04-23 23:19:27.373749991 +0000 UTC m=+1.125872022" Apr 23 23:19:30.788214 kubelet[3427]: I0423 23:19:30.788188 3427 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 23 23:19:30.788900 containerd[1889]: time="2026-04-23T23:19:30.788871369Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 23 23:19:30.789333 kubelet[3427]: I0423 23:19:30.789310 3427 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 23 23:19:31.812786 systemd[1]: Created slice kubepods-besteffort-pode3773903_2d3b_446d_a88f_dbbeab2d536c.slice - libcontainer container kubepods-besteffort-pode3773903_2d3b_446d_a88f_dbbeab2d536c.slice. Apr 23 23:19:31.845140 kubelet[3427]: I0423 23:19:31.845108 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzb75\" (UniqueName: \"kubernetes.io/projected/e3773903-2d3b-446d-a88f-dbbeab2d536c-kube-api-access-rzb75\") pod \"kube-proxy-gw2bt\" (UID: \"e3773903-2d3b-446d-a88f-dbbeab2d536c\") " pod="kube-system/kube-proxy-gw2bt" Apr 23 23:19:31.845140 kubelet[3427]: I0423 23:19:31.845141 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/e3773903-2d3b-446d-a88f-dbbeab2d536c-kube-proxy\") pod \"kube-proxy-gw2bt\" (UID: \"e3773903-2d3b-446d-a88f-dbbeab2d536c\") " pod="kube-system/kube-proxy-gw2bt" Apr 23 23:19:31.845406 kubelet[3427]: I0423 23:19:31.845153 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e3773903-2d3b-446d-a88f-dbbeab2d536c-xtables-lock\") pod \"kube-proxy-gw2bt\" (UID: \"e3773903-2d3b-446d-a88f-dbbeab2d536c\") " pod="kube-system/kube-proxy-gw2bt" Apr 23 23:19:31.845406 kubelet[3427]: I0423 23:19:31.845162 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e3773903-2d3b-446d-a88f-dbbeab2d536c-lib-modules\") pod \"kube-proxy-gw2bt\" (UID: \"e3773903-2d3b-446d-a88f-dbbeab2d536c\") " pod="kube-system/kube-proxy-gw2bt" Apr 23 23:19:31.929297 systemd[1]: Created slice kubepods-besteffort-pod85e21bb1_c8a3_4ac2_94d7_77f58ffb6f1a.slice - libcontainer container kubepods-besteffort-pod85e21bb1_c8a3_4ac2_94d7_77f58ffb6f1a.slice. Apr 23 23:19:31.945557 kubelet[3427]: I0423 23:19:31.945521 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9bxc\" (UniqueName: \"kubernetes.io/projected/85e21bb1-c8a3-4ac2-94d7-77f58ffb6f1a-kube-api-access-q9bxc\") pod \"tigera-operator-6fb8d665dd-27hxn\" (UID: \"85e21bb1-c8a3-4ac2-94d7-77f58ffb6f1a\") " pod="tigera-operator/tigera-operator-6fb8d665dd-27hxn" Apr 23 23:19:31.945647 kubelet[3427]: I0423 23:19:31.945573 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/85e21bb1-c8a3-4ac2-94d7-77f58ffb6f1a-var-lib-calico\") pod \"tigera-operator-6fb8d665dd-27hxn\" (UID: \"85e21bb1-c8a3-4ac2-94d7-77f58ffb6f1a\") " pod="tigera-operator/tigera-operator-6fb8d665dd-27hxn" Apr 23 23:19:32.126020 containerd[1889]: time="2026-04-23T23:19:32.125940141Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gw2bt,Uid:e3773903-2d3b-446d-a88f-dbbeab2d536c,Namespace:kube-system,Attempt:0,}" Apr 23 23:19:32.176957 containerd[1889]: time="2026-04-23T23:19:32.176885413Z" level=info msg="connecting to shim 63f82c04cbf48d213d908214afe36af6470b249086d5c278c84555d86962feda" address="unix:///run/containerd/s/98534f5b0e87697b76661ed2b4c7c1b33c1765c2dc4681f57f2d66e4049d5e3d" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:32.191742 systemd[1]: Started cri-containerd-63f82c04cbf48d213d908214afe36af6470b249086d5c278c84555d86962feda.scope - libcontainer container 63f82c04cbf48d213d908214afe36af6470b249086d5c278c84555d86962feda. Apr 23 23:19:32.210989 containerd[1889]: time="2026-04-23T23:19:32.210921758Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gw2bt,Uid:e3773903-2d3b-446d-a88f-dbbeab2d536c,Namespace:kube-system,Attempt:0,} returns sandbox id \"63f82c04cbf48d213d908214afe36af6470b249086d5c278c84555d86962feda\"" Apr 23 23:19:32.220355 containerd[1889]: time="2026-04-23T23:19:32.220249068Z" level=info msg="CreateContainer within sandbox \"63f82c04cbf48d213d908214afe36af6470b249086d5c278c84555d86962feda\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 23 23:19:32.241544 containerd[1889]: time="2026-04-23T23:19:32.241504487Z" level=info msg="Container 2bd7966eea718b4d63bceb8b0b91ba446867d49132ca08089af1c2b7439a1a28: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:32.242273 containerd[1889]: time="2026-04-23T23:19:32.242058381Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6fb8d665dd-27hxn,Uid:85e21bb1-c8a3-4ac2-94d7-77f58ffb6f1a,Namespace:tigera-operator,Attempt:0,}" Apr 23 23:19:32.267617 containerd[1889]: time="2026-04-23T23:19:32.267313869Z" level=info msg="CreateContainer within sandbox \"63f82c04cbf48d213d908214afe36af6470b249086d5c278c84555d86962feda\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"2bd7966eea718b4d63bceb8b0b91ba446867d49132ca08089af1c2b7439a1a28\"" Apr 23 23:19:32.269275 containerd[1889]: time="2026-04-23T23:19:32.269247321Z" level=info msg="StartContainer for \"2bd7966eea718b4d63bceb8b0b91ba446867d49132ca08089af1c2b7439a1a28\"" Apr 23 23:19:32.273944 containerd[1889]: time="2026-04-23T23:19:32.273918152Z" level=info msg="connecting to shim 2bd7966eea718b4d63bceb8b0b91ba446867d49132ca08089af1c2b7439a1a28" address="unix:///run/containerd/s/98534f5b0e87697b76661ed2b4c7c1b33c1765c2dc4681f57f2d66e4049d5e3d" protocol=ttrpc version=3 Apr 23 23:19:32.288709 systemd[1]: Started cri-containerd-2bd7966eea718b4d63bceb8b0b91ba446867d49132ca08089af1c2b7439a1a28.scope - libcontainer container 2bd7966eea718b4d63bceb8b0b91ba446867d49132ca08089af1c2b7439a1a28. Apr 23 23:19:32.293867 containerd[1889]: time="2026-04-23T23:19:32.293838950Z" level=info msg="connecting to shim 49c0883348e92de368deee83b485877148422d3fb981d3b55e4a974dc7f63e3e" address="unix:///run/containerd/s/8e3ba737a5c9d03719c86cf2d18672258fe82accf2e3c862e8088ef4349dc6ff" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:32.313690 systemd[1]: Started cri-containerd-49c0883348e92de368deee83b485877148422d3fb981d3b55e4a974dc7f63e3e.scope - libcontainer container 49c0883348e92de368deee83b485877148422d3fb981d3b55e4a974dc7f63e3e. Apr 23 23:19:32.354979 containerd[1889]: time="2026-04-23T23:19:32.354911309Z" level=info msg="StartContainer for \"2bd7966eea718b4d63bceb8b0b91ba446867d49132ca08089af1c2b7439a1a28\" returns successfully" Apr 23 23:19:32.357936 containerd[1889]: time="2026-04-23T23:19:32.357810302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6fb8d665dd-27hxn,Uid:85e21bb1-c8a3-4ac2-94d7-77f58ffb6f1a,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"49c0883348e92de368deee83b485877148422d3fb981d3b55e4a974dc7f63e3e\"" Apr 23 23:19:32.359981 containerd[1889]: time="2026-04-23T23:19:32.359923609Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.8\"" Apr 23 23:19:33.391158 kubelet[3427]: I0423 23:19:33.390744 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-gw2bt" podStartSLOduration=2.390730986 podStartE2EDuration="2.390730986s" podCreationTimestamp="2026-04-23 23:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:19:33.370394731 +0000 UTC m=+7.122516762" watchObservedRunningTime="2026-04-23 23:19:33.390730986 +0000 UTC m=+7.142853017" Apr 23 23:19:34.019877 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount189828550.mount: Deactivated successfully. Apr 23 23:19:34.337281 containerd[1889]: time="2026-04-23T23:19:34.337175308Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:34.340137 containerd[1889]: time="2026-04-23T23:19:34.340019752Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.8: active requests=0, bytes read=24868969" Apr 23 23:19:34.344482 containerd[1889]: time="2026-04-23T23:19:34.344310931Z" level=info msg="ImageCreate event name:\"sha256:f37773829212e34063aa0c4c18558c40f2fc7ce0c68e8139b71af2ff71e26790\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:34.351399 containerd[1889]: time="2026-04-23T23:19:34.351376583Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:ce8eeaa3e60794610f3851ee06d296575f7c2efef1e3e1f8ac751a1d87ab979c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:34.352720 containerd[1889]: time="2026-04-23T23:19:34.351792423Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.8\" with image id \"sha256:f37773829212e34063aa0c4c18558c40f2fc7ce0c68e8139b71af2ff71e26790\", repo tag \"quay.io/tigera/operator:v1.40.8\", repo digest \"quay.io/tigera/operator@sha256:ce8eeaa3e60794610f3851ee06d296575f7c2efef1e3e1f8ac751a1d87ab979c\", size \"24864964\" in 1.991844045s" Apr 23 23:19:34.352720 containerd[1889]: time="2026-04-23T23:19:34.351816560Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.8\" returns image reference \"sha256:f37773829212e34063aa0c4c18558c40f2fc7ce0c68e8139b71af2ff71e26790\"" Apr 23 23:19:34.360242 containerd[1889]: time="2026-04-23T23:19:34.360221607Z" level=info msg="CreateContainer within sandbox \"49c0883348e92de368deee83b485877148422d3fb981d3b55e4a974dc7f63e3e\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Apr 23 23:19:34.383015 containerd[1889]: time="2026-04-23T23:19:34.382992344Z" level=info msg="Container 66aada648161c35a1fff30a2e2361de29bb93a5ed09e99c49ab7324b7d83f5a4: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:34.397723 containerd[1889]: time="2026-04-23T23:19:34.397684510Z" level=info msg="CreateContainer within sandbox \"49c0883348e92de368deee83b485877148422d3fb981d3b55e4a974dc7f63e3e\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"66aada648161c35a1fff30a2e2361de29bb93a5ed09e99c49ab7324b7d83f5a4\"" Apr 23 23:19:34.398478 containerd[1889]: time="2026-04-23T23:19:34.398454571Z" level=info msg="StartContainer for \"66aada648161c35a1fff30a2e2361de29bb93a5ed09e99c49ab7324b7d83f5a4\"" Apr 23 23:19:34.399187 containerd[1889]: time="2026-04-23T23:19:34.399162910Z" level=info msg="connecting to shim 66aada648161c35a1fff30a2e2361de29bb93a5ed09e99c49ab7324b7d83f5a4" address="unix:///run/containerd/s/8e3ba737a5c9d03719c86cf2d18672258fe82accf2e3c862e8088ef4349dc6ff" protocol=ttrpc version=3 Apr 23 23:19:34.419710 systemd[1]: Started cri-containerd-66aada648161c35a1fff30a2e2361de29bb93a5ed09e99c49ab7324b7d83f5a4.scope - libcontainer container 66aada648161c35a1fff30a2e2361de29bb93a5ed09e99c49ab7324b7d83f5a4. Apr 23 23:19:34.443382 containerd[1889]: time="2026-04-23T23:19:34.443356268Z" level=info msg="StartContainer for \"66aada648161c35a1fff30a2e2361de29bb93a5ed09e99c49ab7324b7d83f5a4\" returns successfully" Apr 23 23:19:35.373386 kubelet[3427]: I0423 23:19:35.373322 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6fb8d665dd-27hxn" podStartSLOduration=2.379555049 podStartE2EDuration="4.373308543s" podCreationTimestamp="2026-04-23 23:19:31 +0000 UTC" firstStartedPulling="2026-04-23 23:19:32.359325874 +0000 UTC m=+6.111447905" lastFinishedPulling="2026-04-23 23:19:34.353079368 +0000 UTC m=+8.105201399" observedRunningTime="2026-04-23 23:19:35.373186442 +0000 UTC m=+9.125308513" watchObservedRunningTime="2026-04-23 23:19:35.373308543 +0000 UTC m=+9.125430574" Apr 23 23:19:39.442145 sudo[2377]: pam_unix(sudo:session): session closed for user root Apr 23 23:19:39.590968 sshd[2376]: Connection closed by 50.85.169.122 port 53780 Apr 23 23:19:39.591440 sshd-session[2373]: pam_unix(sshd:session): session closed for user core Apr 23 23:19:39.595940 systemd-logind[1871]: Session 9 logged out. Waiting for processes to exit. Apr 23 23:19:39.596765 systemd[1]: sshd@6-10.0.0.13:22-50.85.169.122:53780.service: Deactivated successfully. Apr 23 23:19:39.601870 systemd[1]: session-9.scope: Deactivated successfully. Apr 23 23:19:39.602032 systemd[1]: session-9.scope: Consumed 4.076s CPU time, 221.3M memory peak. Apr 23 23:19:39.604200 systemd-logind[1871]: Removed session 9. Apr 23 23:19:41.553460 systemd[1]: Created slice kubepods-besteffort-pod5da72611_d2f4_4730_b2dd_ae55a874d160.slice - libcontainer container kubepods-besteffort-pod5da72611_d2f4_4730_b2dd_ae55a874d160.slice. Apr 23 23:19:41.608328 kubelet[3427]: I0423 23:19:41.608295 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4ctk\" (UniqueName: \"kubernetes.io/projected/5da72611-d2f4-4730-b2dd-ae55a874d160-kube-api-access-v4ctk\") pod \"calico-typha-7c756b4479-dj9bg\" (UID: \"5da72611-d2f4-4730-b2dd-ae55a874d160\") " pod="calico-system/calico-typha-7c756b4479-dj9bg" Apr 23 23:19:41.608328 kubelet[3427]: I0423 23:19:41.608329 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/5da72611-d2f4-4730-b2dd-ae55a874d160-typha-certs\") pod \"calico-typha-7c756b4479-dj9bg\" (UID: \"5da72611-d2f4-4730-b2dd-ae55a874d160\") " pod="calico-system/calico-typha-7c756b4479-dj9bg" Apr 23 23:19:41.608613 kubelet[3427]: I0423 23:19:41.608342 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5da72611-d2f4-4730-b2dd-ae55a874d160-tigera-ca-bundle\") pod \"calico-typha-7c756b4479-dj9bg\" (UID: \"5da72611-d2f4-4730-b2dd-ae55a874d160\") " pod="calico-system/calico-typha-7c756b4479-dj9bg" Apr 23 23:19:41.632666 systemd[1]: Created slice kubepods-besteffort-pod259f772d_7056_40c7_88e4_91dd32c5c691.slice - libcontainer container kubepods-besteffort-pod259f772d_7056_40c7_88e4_91dd32c5c691.slice. Apr 23 23:19:41.709687 kubelet[3427]: I0423 23:19:41.709654 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tdns\" (UniqueName: \"kubernetes.io/projected/259f772d-7056-40c7-88e4-91dd32c5c691-kube-api-access-2tdns\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709687 kubelet[3427]: I0423 23:19:41.709694 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-cni-log-dir\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709852 kubelet[3427]: I0423 23:19:41.709710 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-cni-net-dir\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709852 kubelet[3427]: I0423 23:19:41.709719 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-sys-fs\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709852 kubelet[3427]: I0423 23:19:41.709728 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/259f772d-7056-40c7-88e4-91dd32c5c691-tigera-ca-bundle\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709852 kubelet[3427]: I0423 23:19:41.709737 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-xtables-lock\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709852 kubelet[3427]: I0423 23:19:41.709788 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/259f772d-7056-40c7-88e4-91dd32c5c691-node-certs\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709936 kubelet[3427]: I0423 23:19:41.709815 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-flexvol-driver-host\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709936 kubelet[3427]: I0423 23:19:41.709828 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-lib-modules\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709936 kubelet[3427]: I0423 23:19:41.709840 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-var-run-calico\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709936 kubelet[3427]: I0423 23:19:41.709861 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-bpffs\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.709936 kubelet[3427]: I0423 23:19:41.709875 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-cni-bin-dir\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.710037 kubelet[3427]: I0423 23:19:41.709884 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-nodeproc\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.710037 kubelet[3427]: I0423 23:19:41.709896 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-var-lib-calico\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.710445 kubelet[3427]: I0423 23:19:41.710371 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/259f772d-7056-40c7-88e4-91dd32c5c691-policysync\") pod \"calico-node-zzgnk\" (UID: \"259f772d-7056-40c7-88e4-91dd32c5c691\") " pod="calico-system/calico-node-zzgnk" Apr 23 23:19:41.740733 kubelet[3427]: E0423 23:19:41.740551 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:41.811545 kubelet[3427]: I0423 23:19:41.811412 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/290a2228-90ec-4821-be49-5abe8a0f00b0-registration-dir\") pod \"csi-node-driver-mlsd9\" (UID: \"290a2228-90ec-4821-be49-5abe8a0f00b0\") " pod="calico-system/csi-node-driver-mlsd9" Apr 23 23:19:41.812035 kubelet[3427]: I0423 23:19:41.811991 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/290a2228-90ec-4821-be49-5abe8a0f00b0-varrun\") pod \"csi-node-driver-mlsd9\" (UID: \"290a2228-90ec-4821-be49-5abe8a0f00b0\") " pod="calico-system/csi-node-driver-mlsd9" Apr 23 23:19:41.813155 kubelet[3427]: I0423 23:19:41.813130 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/290a2228-90ec-4821-be49-5abe8a0f00b0-socket-dir\") pod \"csi-node-driver-mlsd9\" (UID: \"290a2228-90ec-4821-be49-5abe8a0f00b0\") " pod="calico-system/csi-node-driver-mlsd9" Apr 23 23:19:41.813225 kubelet[3427]: I0423 23:19:41.813179 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/290a2228-90ec-4821-be49-5abe8a0f00b0-kubelet-dir\") pod \"csi-node-driver-mlsd9\" (UID: \"290a2228-90ec-4821-be49-5abe8a0f00b0\") " pod="calico-system/csi-node-driver-mlsd9" Apr 23 23:19:41.813225 kubelet[3427]: I0423 23:19:41.813195 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77dm8\" (UniqueName: \"kubernetes.io/projected/290a2228-90ec-4821-be49-5abe8a0f00b0-kube-api-access-77dm8\") pod \"csi-node-driver-mlsd9\" (UID: \"290a2228-90ec-4821-be49-5abe8a0f00b0\") " pod="calico-system/csi-node-driver-mlsd9" Apr 23 23:19:41.819629 kubelet[3427]: E0423 23:19:41.819167 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.819629 kubelet[3427]: W0423 23:19:41.819185 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.819629 kubelet[3427]: E0423 23:19:41.819211 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.819629 kubelet[3427]: E0423 23:19:41.819344 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.819629 kubelet[3427]: W0423 23:19:41.819354 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.819629 kubelet[3427]: E0423 23:19:41.819363 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.819629 kubelet[3427]: E0423 23:19:41.819462 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.819629 kubelet[3427]: W0423 23:19:41.819467 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.819629 kubelet[3427]: E0423 23:19:41.819473 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.819629 kubelet[3427]: E0423 23:19:41.819612 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.819828 kubelet[3427]: W0423 23:19:41.819618 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.819828 kubelet[3427]: E0423 23:19:41.819626 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.819828 kubelet[3427]: E0423 23:19:41.819734 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.819828 kubelet[3427]: W0423 23:19:41.819739 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.819828 kubelet[3427]: E0423 23:19:41.819745 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.820018 kubelet[3427]: E0423 23:19:41.819838 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.820018 kubelet[3427]: W0423 23:19:41.819843 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.820018 kubelet[3427]: E0423 23:19:41.819848 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.820753 kubelet[3427]: E0423 23:19:41.820590 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.820753 kubelet[3427]: W0423 23:19:41.820752 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.820843 kubelet[3427]: E0423 23:19:41.820769 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.821226 kubelet[3427]: E0423 23:19:41.821037 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.821226 kubelet[3427]: W0423 23:19:41.821064 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.821226 kubelet[3427]: E0423 23:19:41.821074 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.821226 kubelet[3427]: E0423 23:19:41.821189 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.821226 kubelet[3427]: W0423 23:19:41.821197 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.821226 kubelet[3427]: E0423 23:19:41.821203 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.821454 kubelet[3427]: E0423 23:19:41.821405 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.821454 kubelet[3427]: W0423 23:19:41.821413 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.821454 kubelet[3427]: E0423 23:19:41.821421 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.823291 kubelet[3427]: E0423 23:19:41.822561 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.823291 kubelet[3427]: W0423 23:19:41.822577 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.823291 kubelet[3427]: E0423 23:19:41.822612 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.823291 kubelet[3427]: E0423 23:19:41.822741 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.823291 kubelet[3427]: W0423 23:19:41.822757 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.823291 kubelet[3427]: E0423 23:19:41.822770 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.823291 kubelet[3427]: E0423 23:19:41.822894 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.823291 kubelet[3427]: W0423 23:19:41.822906 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.823291 kubelet[3427]: E0423 23:19:41.822913 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.823291 kubelet[3427]: E0423 23:19:41.823009 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.823484 kubelet[3427]: W0423 23:19:41.823014 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.823484 kubelet[3427]: E0423 23:19:41.823020 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.823484 kubelet[3427]: E0423 23:19:41.823113 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.823484 kubelet[3427]: W0423 23:19:41.823117 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.823484 kubelet[3427]: E0423 23:19:41.823130 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.823484 kubelet[3427]: E0423 23:19:41.823215 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.823484 kubelet[3427]: W0423 23:19:41.823220 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.823484 kubelet[3427]: E0423 23:19:41.823226 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.824809 kubelet[3427]: E0423 23:19:41.824762 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.824809 kubelet[3427]: W0423 23:19:41.824803 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.824884 kubelet[3427]: E0423 23:19:41.824816 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.824996 kubelet[3427]: E0423 23:19:41.824981 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.824996 kubelet[3427]: W0423 23:19:41.824991 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.825040 kubelet[3427]: E0423 23:19:41.825015 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.825148 kubelet[3427]: E0423 23:19:41.825134 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.825148 kubelet[3427]: W0423 23:19:41.825143 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.825203 kubelet[3427]: E0423 23:19:41.825178 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.825479 kubelet[3427]: E0423 23:19:41.825454 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.825532 kubelet[3427]: W0423 23:19:41.825483 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.825532 kubelet[3427]: E0423 23:19:41.825494 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.825832 kubelet[3427]: E0423 23:19:41.825810 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.826196 kubelet[3427]: W0423 23:19:41.825903 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.826196 kubelet[3427]: E0423 23:19:41.825920 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.826264 kubelet[3427]: E0423 23:19:41.826256 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.826281 kubelet[3427]: W0423 23:19:41.826265 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.826281 kubelet[3427]: E0423 23:19:41.826276 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.826549 kubelet[3427]: E0423 23:19:41.826499 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.826549 kubelet[3427]: W0423 23:19:41.826543 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.826549 kubelet[3427]: E0423 23:19:41.826552 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.826735 kubelet[3427]: E0423 23:19:41.826719 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.826735 kubelet[3427]: W0423 23:19:41.826734 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.826778 kubelet[3427]: E0423 23:19:41.826742 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.831966 kubelet[3427]: E0423 23:19:41.831943 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.831966 kubelet[3427]: W0423 23:19:41.831961 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.832052 kubelet[3427]: E0423 23:19:41.831973 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.864975 containerd[1889]: time="2026-04-23T23:19:41.864699070Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c756b4479-dj9bg,Uid:5da72611-d2f4-4730-b2dd-ae55a874d160,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:41.903461 containerd[1889]: time="2026-04-23T23:19:41.903383096Z" level=info msg="connecting to shim 29eacf10ef4d5e7f1e83e3024683cbc8c31bc5f44405769ac76e6a0be11a0f80" address="unix:///run/containerd/s/6c3eb3c7b06ae004159188147e8b0b4d20552e65f1e48ff4da5ce7b526ae8b06" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:41.914313 kubelet[3427]: E0423 23:19:41.914284 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.914313 kubelet[3427]: W0423 23:19:41.914298 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.914313 kubelet[3427]: E0423 23:19:41.914309 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.915746 kubelet[3427]: E0423 23:19:41.915732 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.915953 kubelet[3427]: W0423 23:19:41.915851 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.915953 kubelet[3427]: E0423 23:19:41.915866 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.916130 kubelet[3427]: E0423 23:19:41.916117 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.916311 kubelet[3427]: W0423 23:19:41.916190 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.916311 kubelet[3427]: E0423 23:19:41.916214 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.916588 kubelet[3427]: E0423 23:19:41.916483 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.916588 kubelet[3427]: W0423 23:19:41.916494 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.916588 kubelet[3427]: E0423 23:19:41.916504 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.916914 kubelet[3427]: E0423 23:19:41.916893 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.917015 kubelet[3427]: W0423 23:19:41.916965 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.917015 kubelet[3427]: E0423 23:19:41.917002 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.917406 kubelet[3427]: E0423 23:19:41.917318 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.917406 kubelet[3427]: W0423 23:19:41.917330 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.917406 kubelet[3427]: E0423 23:19:41.917339 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.918071 kubelet[3427]: E0423 23:19:41.918057 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.918268 kubelet[3427]: W0423 23:19:41.918155 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.918268 kubelet[3427]: E0423 23:19:41.918171 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.918490 kubelet[3427]: E0423 23:19:41.918443 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.918490 kubelet[3427]: W0423 23:19:41.918453 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.918490 kubelet[3427]: E0423 23:19:41.918462 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.918897 kubelet[3427]: E0423 23:19:41.918821 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.918897 kubelet[3427]: W0423 23:19:41.918833 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.918897 kubelet[3427]: E0423 23:19:41.918842 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.919386 kubelet[3427]: E0423 23:19:41.919294 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.919615 kubelet[3427]: W0423 23:19:41.919453 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.919615 kubelet[3427]: E0423 23:19:41.919470 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.920168 kubelet[3427]: E0423 23:19:41.920153 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.920460 kubelet[3427]: W0423 23:19:41.920313 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.920460 kubelet[3427]: E0423 23:19:41.920332 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.920588 kubelet[3427]: E0423 23:19:41.920578 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.920666 kubelet[3427]: W0423 23:19:41.920655 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.920716 kubelet[3427]: E0423 23:19:41.920707 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.921160 kubelet[3427]: E0423 23:19:41.921125 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.921486 kubelet[3427]: W0423 23:19:41.921412 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.921486 kubelet[3427]: E0423 23:19:41.921437 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.921793 systemd[1]: Started cri-containerd-29eacf10ef4d5e7f1e83e3024683cbc8c31bc5f44405769ac76e6a0be11a0f80.scope - libcontainer container 29eacf10ef4d5e7f1e83e3024683cbc8c31bc5f44405769ac76e6a0be11a0f80. Apr 23 23:19:41.922392 kubelet[3427]: E0423 23:19:41.922378 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.922614 kubelet[3427]: W0423 23:19:41.922440 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.922614 kubelet[3427]: E0423 23:19:41.922455 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.922943 kubelet[3427]: E0423 23:19:41.922872 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.922943 kubelet[3427]: W0423 23:19:41.922886 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.922943 kubelet[3427]: E0423 23:19:41.922895 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.923353 kubelet[3427]: E0423 23:19:41.923340 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.923809 kubelet[3427]: W0423 23:19:41.923396 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.923809 kubelet[3427]: E0423 23:19:41.923410 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.924264 kubelet[3427]: E0423 23:19:41.924126 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.924264 kubelet[3427]: W0423 23:19:41.924194 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.924264 kubelet[3427]: E0423 23:19:41.924209 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.924690 kubelet[3427]: E0423 23:19:41.924589 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.924690 kubelet[3427]: W0423 23:19:41.924620 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.924690 kubelet[3427]: E0423 23:19:41.924633 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.925483 kubelet[3427]: E0423 23:19:41.925468 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.925629 kubelet[3427]: W0423 23:19:41.925544 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.925629 kubelet[3427]: E0423 23:19:41.925558 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.925851 kubelet[3427]: E0423 23:19:41.925840 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.925984 kubelet[3427]: W0423 23:19:41.925916 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.925984 kubelet[3427]: E0423 23:19:41.925931 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.926393 kubelet[3427]: E0423 23:19:41.926247 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.926393 kubelet[3427]: W0423 23:19:41.926312 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.926393 kubelet[3427]: E0423 23:19:41.926326 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.927015 kubelet[3427]: E0423 23:19:41.927000 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.927260 kubelet[3427]: W0423 23:19:41.927067 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.927260 kubelet[3427]: E0423 23:19:41.927080 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.927633 kubelet[3427]: E0423 23:19:41.927618 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.927808 kubelet[3427]: W0423 23:19:41.927794 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.928423 kubelet[3427]: E0423 23:19:41.928343 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.928665 kubelet[3427]: E0423 23:19:41.928651 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.928836 kubelet[3427]: W0423 23:19:41.928744 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.928836 kubelet[3427]: E0423 23:19:41.928768 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.929062 kubelet[3427]: E0423 23:19:41.929047 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.929383 kubelet[3427]: W0423 23:19:41.929114 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.929383 kubelet[3427]: E0423 23:19:41.929128 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.939830 kubelet[3427]: E0423 23:19:41.939811 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:41.939830 kubelet[3427]: W0423 23:19:41.939827 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:41.939933 kubelet[3427]: E0423 23:19:41.939837 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:41.943647 containerd[1889]: time="2026-04-23T23:19:41.943619274Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-zzgnk,Uid:259f772d-7056-40c7-88e4-91dd32c5c691,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:41.995317 containerd[1889]: time="2026-04-23T23:19:41.995242485Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c756b4479-dj9bg,Uid:5da72611-d2f4-4730-b2dd-ae55a874d160,Namespace:calico-system,Attempt:0,} returns sandbox id \"29eacf10ef4d5e7f1e83e3024683cbc8c31bc5f44405769ac76e6a0be11a0f80\"" Apr 23 23:19:41.998053 containerd[1889]: time="2026-04-23T23:19:41.997897582Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.5\"" Apr 23 23:19:42.020912 containerd[1889]: time="2026-04-23T23:19:42.020809006Z" level=info msg="connecting to shim efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10" address="unix:///run/containerd/s/53309e2d30f48cf732877c6e74c1f75ec93eb58cc38f7a162cc072a81364c299" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:42.043703 systemd[1]: Started cri-containerd-efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10.scope - libcontainer container efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10. Apr 23 23:19:42.067273 containerd[1889]: time="2026-04-23T23:19:42.067190977Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-zzgnk,Uid:259f772d-7056-40c7-88e4-91dd32c5c691,Namespace:calico-system,Attempt:0,} returns sandbox id \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\"" Apr 23 23:19:43.184977 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount439905952.mount: Deactivated successfully. Apr 23 23:19:43.316919 kubelet[3427]: E0423 23:19:43.316872 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:43.627553 containerd[1889]: time="2026-04-23T23:19:43.627513208Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:43.631334 containerd[1889]: time="2026-04-23T23:19:43.631308083Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.5: active requests=0, bytes read=32841445" Apr 23 23:19:43.634106 containerd[1889]: time="2026-04-23T23:19:43.634080209Z" level=info msg="ImageCreate event name:\"sha256:265c145eea96693e7abfe97a68dee913c8e656947f5708c28e4e866d3809b4c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:43.639035 containerd[1889]: time="2026-04-23T23:19:43.638617047Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:76afd8f80569b3bf783991ce5348294319cefa6d6cca127710d0e068096048a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:43.639035 containerd[1889]: time="2026-04-23T23:19:43.638944779Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.5\" with image id \"sha256:265c145eea96693e7abfe97a68dee913c8e656947f5708c28e4e866d3809b4c9\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:76afd8f80569b3bf783991ce5348294319cefa6d6cca127710d0e068096048a6\", size \"32841299\" in 1.640795228s" Apr 23 23:19:43.639035 containerd[1889]: time="2026-04-23T23:19:43.638967556Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.5\" returns image reference \"sha256:265c145eea96693e7abfe97a68dee913c8e656947f5708c28e4e866d3809b4c9\"" Apr 23 23:19:43.640311 containerd[1889]: time="2026-04-23T23:19:43.640151263Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\"" Apr 23 23:19:43.655396 containerd[1889]: time="2026-04-23T23:19:43.655366453Z" level=info msg="CreateContainer within sandbox \"29eacf10ef4d5e7f1e83e3024683cbc8c31bc5f44405769ac76e6a0be11a0f80\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Apr 23 23:19:43.673767 containerd[1889]: time="2026-04-23T23:19:43.673708181Z" level=info msg="Container e76d6800383a1e7003f8edfa09179966135ba5fa4538acaa1ef8cc99aecc5711: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:43.690644 containerd[1889]: time="2026-04-23T23:19:43.690586543Z" level=info msg="CreateContainer within sandbox \"29eacf10ef4d5e7f1e83e3024683cbc8c31bc5f44405769ac76e6a0be11a0f80\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"e76d6800383a1e7003f8edfa09179966135ba5fa4538acaa1ef8cc99aecc5711\"" Apr 23 23:19:43.691718 containerd[1889]: time="2026-04-23T23:19:43.691699664Z" level=info msg="StartContainer for \"e76d6800383a1e7003f8edfa09179966135ba5fa4538acaa1ef8cc99aecc5711\"" Apr 23 23:19:43.692791 containerd[1889]: time="2026-04-23T23:19:43.692718789Z" level=info msg="connecting to shim e76d6800383a1e7003f8edfa09179966135ba5fa4538acaa1ef8cc99aecc5711" address="unix:///run/containerd/s/6c3eb3c7b06ae004159188147e8b0b4d20552e65f1e48ff4da5ce7b526ae8b06" protocol=ttrpc version=3 Apr 23 23:19:43.712706 systemd[1]: Started cri-containerd-e76d6800383a1e7003f8edfa09179966135ba5fa4538acaa1ef8cc99aecc5711.scope - libcontainer container e76d6800383a1e7003f8edfa09179966135ba5fa4538acaa1ef8cc99aecc5711. Apr 23 23:19:43.750813 containerd[1889]: time="2026-04-23T23:19:43.750745315Z" level=info msg="StartContainer for \"e76d6800383a1e7003f8edfa09179966135ba5fa4538acaa1ef8cc99aecc5711\" returns successfully" Apr 23 23:19:44.404513 kubelet[3427]: E0423 23:19:44.404483 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.404513 kubelet[3427]: W0423 23:19:44.404503 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.404513 kubelet[3427]: E0423 23:19:44.404519 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.404872 kubelet[3427]: E0423 23:19:44.404651 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.404872 kubelet[3427]: W0423 23:19:44.404658 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.404872 kubelet[3427]: E0423 23:19:44.404683 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.404872 kubelet[3427]: E0423 23:19:44.404782 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.404872 kubelet[3427]: W0423 23:19:44.404786 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.404872 kubelet[3427]: E0423 23:19:44.404791 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.404872 kubelet[3427]: E0423 23:19:44.404876 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.404974 kubelet[3427]: W0423 23:19:44.404880 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.404974 kubelet[3427]: E0423 23:19:44.404885 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405003 kubelet[3427]: E0423 23:19:44.404977 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405003 kubelet[3427]: W0423 23:19:44.404981 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405003 kubelet[3427]: E0423 23:19:44.404986 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405078 kubelet[3427]: E0423 23:19:44.405061 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405078 kubelet[3427]: W0423 23:19:44.405072 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405078 kubelet[3427]: E0423 23:19:44.405077 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405167 kubelet[3427]: E0423 23:19:44.405151 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405167 kubelet[3427]: W0423 23:19:44.405161 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405167 kubelet[3427]: E0423 23:19:44.405165 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405257 kubelet[3427]: E0423 23:19:44.405244 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405257 kubelet[3427]: W0423 23:19:44.405252 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405257 kubelet[3427]: E0423 23:19:44.405257 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405358 kubelet[3427]: E0423 23:19:44.405345 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405358 kubelet[3427]: W0423 23:19:44.405354 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405358 kubelet[3427]: E0423 23:19:44.405361 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405445 kubelet[3427]: E0423 23:19:44.405437 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405445 kubelet[3427]: W0423 23:19:44.405443 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405475 kubelet[3427]: E0423 23:19:44.405449 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405535 kubelet[3427]: E0423 23:19:44.405522 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405535 kubelet[3427]: W0423 23:19:44.405530 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405535 kubelet[3427]: E0423 23:19:44.405535 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405634 kubelet[3427]: E0423 23:19:44.405621 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405634 kubelet[3427]: W0423 23:19:44.405630 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405672 kubelet[3427]: E0423 23:19:44.405635 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405738 kubelet[3427]: E0423 23:19:44.405725 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405738 kubelet[3427]: W0423 23:19:44.405734 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405783 kubelet[3427]: E0423 23:19:44.405739 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405838 kubelet[3427]: E0423 23:19:44.405820 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405838 kubelet[3427]: W0423 23:19:44.405828 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405838 kubelet[3427]: E0423 23:19:44.405832 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.405915 kubelet[3427]: E0423 23:19:44.405906 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.405915 kubelet[3427]: W0423 23:19:44.405913 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.405945 kubelet[3427]: E0423 23:19:44.405918 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.432835 kubelet[3427]: E0423 23:19:44.432765 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.432835 kubelet[3427]: W0423 23:19:44.432778 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.432835 kubelet[3427]: E0423 23:19:44.432788 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.433116 kubelet[3427]: E0423 23:19:44.433097 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.433216 kubelet[3427]: W0423 23:19:44.433109 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.433216 kubelet[3427]: E0423 23:19:44.433176 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.433342 kubelet[3427]: E0423 23:19:44.433319 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.433342 kubelet[3427]: W0423 23:19:44.433332 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.433439 kubelet[3427]: E0423 23:19:44.433342 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.433439 kubelet[3427]: E0423 23:19:44.433446 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.433514 kubelet[3427]: W0423 23:19:44.433452 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.433514 kubelet[3427]: E0423 23:19:44.433458 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.433667 kubelet[3427]: E0423 23:19:44.433542 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.433667 kubelet[3427]: W0423 23:19:44.433546 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.433667 kubelet[3427]: E0423 23:19:44.433552 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.433756 kubelet[3427]: E0423 23:19:44.433698 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.433756 kubelet[3427]: W0423 23:19:44.433705 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.433756 kubelet[3427]: E0423 23:19:44.433711 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.434023 kubelet[3427]: E0423 23:19:44.434012 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.434092 kubelet[3427]: W0423 23:19:44.434082 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.434131 kubelet[3427]: E0423 23:19:44.434122 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.434292 kubelet[3427]: E0423 23:19:44.434279 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.434292 kubelet[3427]: W0423 23:19:44.434291 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.434374 kubelet[3427]: E0423 23:19:44.434300 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.434431 kubelet[3427]: E0423 23:19:44.434399 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.434431 kubelet[3427]: W0423 23:19:44.434403 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.434431 kubelet[3427]: E0423 23:19:44.434409 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.434524 kubelet[3427]: E0423 23:19:44.434489 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.434524 kubelet[3427]: W0423 23:19:44.434493 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.434524 kubelet[3427]: E0423 23:19:44.434498 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.434640 kubelet[3427]: E0423 23:19:44.434609 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.434640 kubelet[3427]: W0423 23:19:44.434614 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.434640 kubelet[3427]: E0423 23:19:44.434621 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.434913 kubelet[3427]: E0423 23:19:44.434901 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.434983 kubelet[3427]: W0423 23:19:44.434973 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.435111 kubelet[3427]: E0423 23:19:44.435030 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.435295 kubelet[3427]: E0423 23:19:44.435284 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.435358 kubelet[3427]: W0423 23:19:44.435348 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.435454 kubelet[3427]: E0423 23:19:44.435402 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.435958 kubelet[3427]: E0423 23:19:44.435933 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.436116 kubelet[3427]: W0423 23:19:44.436050 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.436116 kubelet[3427]: E0423 23:19:44.436066 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.436377 kubelet[3427]: E0423 23:19:44.436343 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.436377 kubelet[3427]: W0423 23:19:44.436354 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.436377 kubelet[3427]: E0423 23:19:44.436364 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.436682 kubelet[3427]: E0423 23:19:44.436619 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.436682 kubelet[3427]: W0423 23:19:44.436630 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.436682 kubelet[3427]: E0423 23:19:44.436639 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.436962 kubelet[3427]: E0423 23:19:44.436925 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.436962 kubelet[3427]: W0423 23:19:44.436936 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.436962 kubelet[3427]: E0423 23:19:44.436946 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.437381 kubelet[3427]: E0423 23:19:44.437331 3427 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 23 23:19:44.437381 kubelet[3427]: W0423 23:19:44.437350 3427 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 23 23:19:44.437381 kubelet[3427]: E0423 23:19:44.437358 3427 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 23 23:19:44.819146 containerd[1889]: time="2026-04-23T23:19:44.818887937Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:44.822999 containerd[1889]: time="2026-04-23T23:19:44.822970447Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5: active requests=0, bytes read=4404646" Apr 23 23:19:44.825867 containerd[1889]: time="2026-04-23T23:19:44.825830775Z" level=info msg="ImageCreate event name:\"sha256:3867b4c2eaa3321472d76c87dc2b4f8d6cdd45473f2138098e7ef206bc16d421\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:44.831088 containerd[1889]: time="2026-04-23T23:19:44.831047399Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:df00fee6895ac073066d91243f29733e71f479317cacef49d50c244bb2d21ea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:44.832006 containerd[1889]: time="2026-04-23T23:19:44.831916886Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\" with image id \"sha256:3867b4c2eaa3321472d76c87dc2b4f8d6cdd45473f2138098e7ef206bc16d421\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:df00fee6895ac073066d91243f29733e71f479317cacef49d50c244bb2d21ea1\", size \"6980245\" in 1.19174043s" Apr 23 23:19:44.832006 containerd[1889]: time="2026-04-23T23:19:44.831944487Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\" returns image reference \"sha256:3867b4c2eaa3321472d76c87dc2b4f8d6cdd45473f2138098e7ef206bc16d421\"" Apr 23 23:19:44.843541 containerd[1889]: time="2026-04-23T23:19:44.843519920Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Apr 23 23:19:44.865355 containerd[1889]: time="2026-04-23T23:19:44.864699176Z" level=info msg="Container 07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:44.881854 containerd[1889]: time="2026-04-23T23:19:44.881826067Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a\"" Apr 23 23:19:44.883160 containerd[1889]: time="2026-04-23T23:19:44.883144371Z" level=info msg="StartContainer for \"07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a\"" Apr 23 23:19:44.884274 containerd[1889]: time="2026-04-23T23:19:44.884230707Z" level=info msg="connecting to shim 07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a" address="unix:///run/containerd/s/53309e2d30f48cf732877c6e74c1f75ec93eb58cc38f7a162cc072a81364c299" protocol=ttrpc version=3 Apr 23 23:19:44.899706 systemd[1]: Started cri-containerd-07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a.scope - libcontainer container 07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a. Apr 23 23:19:44.951567 containerd[1889]: time="2026-04-23T23:19:44.951492340Z" level=info msg="StartContainer for \"07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a\" returns successfully" Apr 23 23:19:44.953996 systemd[1]: cri-containerd-07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a.scope: Deactivated successfully. Apr 23 23:19:44.958620 containerd[1889]: time="2026-04-23T23:19:44.958528053Z" level=info msg="received container exit event container_id:\"07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a\" id:\"07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a\" pid:4067 exited_at:{seconds:1776986384 nanos:958101790}" Apr 23 23:19:44.972257 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-07cfe2028e7f77f85edca74c997c09305539031ea16498d4a07be1e9e356b85a-rootfs.mount: Deactivated successfully. Apr 23 23:19:45.316468 kubelet[3427]: E0423 23:19:45.316425 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:45.384482 kubelet[3427]: I0423 23:19:45.384423 3427 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 23 23:19:45.400253 kubelet[3427]: I0423 23:19:45.400206 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7c756b4479-dj9bg" podStartSLOduration=2.758153025 podStartE2EDuration="4.400105183s" podCreationTimestamp="2026-04-23 23:19:41 +0000 UTC" firstStartedPulling="2026-04-23 23:19:41.997681758 +0000 UTC m=+15.749803789" lastFinishedPulling="2026-04-23 23:19:43.639633916 +0000 UTC m=+17.391755947" observedRunningTime="2026-04-23 23:19:44.39328104 +0000 UTC m=+18.145403071" watchObservedRunningTime="2026-04-23 23:19:45.400105183 +0000 UTC m=+19.152227214" Apr 23 23:19:46.388607 containerd[1889]: time="2026-04-23T23:19:46.388501292Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.5\"" Apr 23 23:19:47.316889 kubelet[3427]: E0423 23:19:47.316849 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:49.316947 kubelet[3427]: E0423 23:19:49.316780 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:49.995495 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount887740945.mount: Deactivated successfully. Apr 23 23:19:50.649110 containerd[1889]: time="2026-04-23T23:19:50.648643314Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:50.651304 containerd[1889]: time="2026-04-23T23:19:50.651282989Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.5: active requests=0, bytes read=153029581" Apr 23 23:19:50.654061 containerd[1889]: time="2026-04-23T23:19:50.654038668Z" level=info msg="ImageCreate event name:\"sha256:5a8f90ba0ad45873b37c9c512d6391f35086ced5c27f20cfc5c45f777f9941b3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:50.658717 containerd[1889]: time="2026-04-23T23:19:50.658390343Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e2426b97a645ed620e0f4035d594f2f3344b0547cd3dc3458f45e06d5cebdad7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:50.658717 containerd[1889]: time="2026-04-23T23:19:50.658640473Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.5\" with image id \"sha256:5a8f90ba0ad45873b37c9c512d6391f35086ced5c27f20cfc5c45f777f9941b3\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e2426b97a645ed620e0f4035d594f2f3344b0547cd3dc3458f45e06d5cebdad7\", size \"153029443\" in 4.269649611s" Apr 23 23:19:50.658717 containerd[1889]: time="2026-04-23T23:19:50.658659657Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.5\" returns image reference \"sha256:5a8f90ba0ad45873b37c9c512d6391f35086ced5c27f20cfc5c45f777f9941b3\"" Apr 23 23:19:50.666407 containerd[1889]: time="2026-04-23T23:19:50.666372283Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Apr 23 23:19:50.686900 containerd[1889]: time="2026-04-23T23:19:50.686847091Z" level=info msg="Container 88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:50.689913 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount757388084.mount: Deactivated successfully. Apr 23 23:19:50.706892 containerd[1889]: time="2026-04-23T23:19:50.706842194Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e\"" Apr 23 23:19:50.707359 containerd[1889]: time="2026-04-23T23:19:50.707322972Z" level=info msg="StartContainer for \"88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e\"" Apr 23 23:19:50.709216 containerd[1889]: time="2026-04-23T23:19:50.709186466Z" level=info msg="connecting to shim 88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e" address="unix:///run/containerd/s/53309e2d30f48cf732877c6e74c1f75ec93eb58cc38f7a162cc072a81364c299" protocol=ttrpc version=3 Apr 23 23:19:50.727711 systemd[1]: Started cri-containerd-88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e.scope - libcontainer container 88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e. Apr 23 23:19:50.773645 containerd[1889]: time="2026-04-23T23:19:50.773615748Z" level=info msg="StartContainer for \"88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e\" returns successfully" Apr 23 23:19:50.805315 systemd[1]: cri-containerd-88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e.scope: Deactivated successfully. Apr 23 23:19:50.808932 containerd[1889]: time="2026-04-23T23:19:50.808817981Z" level=info msg="received container exit event container_id:\"88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e\" id:\"88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e\" pid:4122 exited_at:{seconds:1776986390 nanos:807522397}" Apr 23 23:19:50.994310 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-88facadc6b3bcfe0f7bda38c134ccb9c85947514743d0628a8e30f08a0b2fb6e-rootfs.mount: Deactivated successfully. Apr 23 23:19:51.316663 kubelet[3427]: E0423 23:19:51.316389 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:52.417078 containerd[1889]: time="2026-04-23T23:19:52.416389331Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.5\"" Apr 23 23:19:53.316682 kubelet[3427]: E0423 23:19:53.316641 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:54.611477 containerd[1889]: time="2026-04-23T23:19:54.611433100Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:54.614640 containerd[1889]: time="2026-04-23T23:19:54.614610835Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.5: active requests=0, bytes read=62266008" Apr 23 23:19:54.618131 containerd[1889]: time="2026-04-23T23:19:54.618100654Z" level=info msg="ImageCreate event name:\"sha256:0636f5f0fe5e716fd01c674abaaef326193e41f0291d3a9b0ce572a82500c211\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:54.622159 containerd[1889]: time="2026-04-23T23:19:54.622128349Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:ea8a6b721af629c1dab2e1559b93cd843d9a4b640726115380fc23cf47e83232\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:19:54.622605 containerd[1889]: time="2026-04-23T23:19:54.622574566Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.5\" with image id \"sha256:0636f5f0fe5e716fd01c674abaaef326193e41f0291d3a9b0ce572a82500c211\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:ea8a6b721af629c1dab2e1559b93cd843d9a4b640726115380fc23cf47e83232\", size \"64841647\" in 2.205607949s" Apr 23 23:19:54.622774 containerd[1889]: time="2026-04-23T23:19:54.622756477Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.5\" returns image reference \"sha256:0636f5f0fe5e716fd01c674abaaef326193e41f0291d3a9b0ce572a82500c211\"" Apr 23 23:19:54.630892 containerd[1889]: time="2026-04-23T23:19:54.630864117Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Apr 23 23:19:54.657408 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4106166324.mount: Deactivated successfully. Apr 23 23:19:54.658689 containerd[1889]: time="2026-04-23T23:19:54.657868531Z" level=info msg="Container 4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:54.675852 containerd[1889]: time="2026-04-23T23:19:54.675820837Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8\"" Apr 23 23:19:54.676277 containerd[1889]: time="2026-04-23T23:19:54.676148641Z" level=info msg="StartContainer for \"4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8\"" Apr 23 23:19:54.677344 containerd[1889]: time="2026-04-23T23:19:54.677319389Z" level=info msg="connecting to shim 4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8" address="unix:///run/containerd/s/53309e2d30f48cf732877c6e74c1f75ec93eb58cc38f7a162cc072a81364c299" protocol=ttrpc version=3 Apr 23 23:19:54.693713 systemd[1]: Started cri-containerd-4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8.scope - libcontainer container 4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8. Apr 23 23:19:54.748696 containerd[1889]: time="2026-04-23T23:19:54.748653890Z" level=info msg="StartContainer for \"4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8\" returns successfully" Apr 23 23:19:55.317010 kubelet[3427]: E0423 23:19:55.316962 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:55.897203 containerd[1889]: time="2026-04-23T23:19:55.897162828Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 23 23:19:55.899193 systemd[1]: cri-containerd-4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8.scope: Deactivated successfully. Apr 23 23:19:55.899436 systemd[1]: cri-containerd-4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8.scope: Consumed 301ms CPU time, 194M memory peak, 165.6M written to disk. Apr 23 23:19:55.900979 containerd[1889]: time="2026-04-23T23:19:55.900948674Z" level=info msg="received container exit event container_id:\"4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8\" id:\"4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8\" pid:4183 exited_at:{seconds:1776986395 nanos:900674664}" Apr 23 23:19:55.913172 kubelet[3427]: I0423 23:19:55.913153 3427 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Apr 23 23:19:55.920474 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8-rootfs.mount: Deactivated successfully. Apr 23 23:19:56.801775 systemd[1]: Created slice kubepods-besteffort-pod043b430d_28cd_4c76_9d72_4b30392409aa.slice - libcontainer container kubepods-besteffort-pod043b430d_28cd_4c76_9d72_4b30392409aa.slice. Apr 23 23:19:56.808297 containerd[1889]: time="2026-04-23T23:19:56.808245479Z" level=error msg="collecting metrics for 4687c46cdef187072a64b5e8b65b4f4a2c3c1563158532426a3a49c36befbca8" error="ttrpc: closed" Apr 23 23:19:56.810400 systemd[1]: Created slice kubepods-besteffort-pod290a2228_90ec_4821_be49_5abe8a0f00b0.slice - libcontainer container kubepods-besteffort-pod290a2228_90ec_4821_be49_5abe8a0f00b0.slice. Apr 23 23:19:56.819470 systemd[1]: Created slice kubepods-burstable-pod9bd25f44_e1d9_4f63_84fb_09a335c9c974.slice - libcontainer container kubepods-burstable-pod9bd25f44_e1d9_4f63_84fb_09a335c9c974.slice. Apr 23 23:19:56.826315 systemd[1]: Created slice kubepods-burstable-podf142fda8_6ffd_4418_b724_d55c78399964.slice - libcontainer container kubepods-burstable-podf142fda8_6ffd_4418_b724_d55c78399964.slice. Apr 23 23:19:56.832573 containerd[1889]: time="2026-04-23T23:19:56.832545055Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mlsd9,Uid:290a2228-90ec-4821-be49-5abe8a0f00b0,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:56.838615 systemd[1]: Created slice kubepods-besteffort-podb94f8cb1_0ece_4e1e_a59c_df0ce7bed609.slice - libcontainer container kubepods-besteffort-podb94f8cb1_0ece_4e1e_a59c_df0ce7bed609.slice. Apr 23 23:19:56.852968 systemd[1]: Created slice kubepods-besteffort-podfb79fb64_b4a9_40ad_9ed6_8c74737a8635.slice - libcontainer container kubepods-besteffort-podfb79fb64_b4a9_40ad_9ed6_8c74737a8635.slice. Apr 23 23:19:56.866931 systemd[1]: Created slice kubepods-besteffort-pod9035f2c7_dbfc_4199_a5d8_b57d671b43bb.slice - libcontainer container kubepods-besteffort-pod9035f2c7_dbfc_4199_a5d8_b57d671b43bb.slice. Apr 23 23:19:56.872497 systemd[1]: Created slice kubepods-besteffort-pod48f3e795_c1d9_4588_9444_c43e72073baf.slice - libcontainer container kubepods-besteffort-pod48f3e795_c1d9_4588_9444_c43e72073baf.slice. Apr 23 23:19:56.899505 containerd[1889]: time="2026-04-23T23:19:56.898257217Z" level=error msg="Failed to destroy network for sandbox \"a22f1c8b373c4c77d95a1e23cba5252760d706ae5f3ac751cf5344746a934b9c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:56.899487 systemd[1]: run-netns-cni\x2d57db461c\x2d0bb3\x2da5c0\x2debfd\x2dc545ec280f97.mount: Deactivated successfully. Apr 23 23:19:56.905176 containerd[1889]: time="2026-04-23T23:19:56.905114786Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mlsd9,Uid:290a2228-90ec-4821-be49-5abe8a0f00b0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a22f1c8b373c4c77d95a1e23cba5252760d706ae5f3ac751cf5344746a934b9c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:56.905407 kubelet[3427]: E0423 23:19:56.905375 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a22f1c8b373c4c77d95a1e23cba5252760d706ae5f3ac751cf5344746a934b9c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:56.905844 kubelet[3427]: E0423 23:19:56.905426 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a22f1c8b373c4c77d95a1e23cba5252760d706ae5f3ac751cf5344746a934b9c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-mlsd9" Apr 23 23:19:56.905844 kubelet[3427]: E0423 23:19:56.905438 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a22f1c8b373c4c77d95a1e23cba5252760d706ae5f3ac751cf5344746a934b9c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-mlsd9" Apr 23 23:19:56.905844 kubelet[3427]: E0423 23:19:56.905478 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-mlsd9_calico-system(290a2228-90ec-4821-be49-5abe8a0f00b0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-mlsd9_calico-system(290a2228-90ec-4821-be49-5abe8a0f00b0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a22f1c8b373c4c77d95a1e23cba5252760d706ae5f3ac751cf5344746a934b9c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-mlsd9" podUID="290a2228-90ec-4821-be49-5abe8a0f00b0" Apr 23 23:19:56.912160 kubelet[3427]: I0423 23:19:56.912113 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-backend-key-pair\") pod \"whisker-bd885f7c6-5mtnn\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " pod="calico-system/whisker-bd885f7c6-5mtnn" Apr 23 23:19:56.912160 kubelet[3427]: I0423 23:19:56.912138 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bd25f44-e1d9-4f63-84fb-09a335c9c974-config-volume\") pod \"coredns-66bc5c9577-zc5vb\" (UID: \"9bd25f44-e1d9-4f63-84fb-09a335c9c974\") " pod="kube-system/coredns-66bc5c9577-zc5vb" Apr 23 23:19:56.912259 kubelet[3427]: I0423 23:19:56.912227 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-nginx-config\") pod \"whisker-bd885f7c6-5mtnn\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " pod="calico-system/whisker-bd885f7c6-5mtnn" Apr 23 23:19:56.912259 kubelet[3427]: I0423 23:19:56.912249 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/043b430d-28cd-4c76-9d72-4b30392409aa-tigera-ca-bundle\") pod \"calico-kube-controllers-f94c58f4-b8pwc\" (UID: \"043b430d-28cd-4c76-9d72-4b30392409aa\") " pod="calico-system/calico-kube-controllers-f94c58f4-b8pwc" Apr 23 23:19:56.912291 kubelet[3427]: I0423 23:19:56.912264 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvd2d\" (UniqueName: \"kubernetes.io/projected/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-kube-api-access-lvd2d\") pod \"whisker-bd885f7c6-5mtnn\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " pod="calico-system/whisker-bd885f7c6-5mtnn" Apr 23 23:19:56.912291 kubelet[3427]: I0423 23:19:56.912274 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msr62\" (UniqueName: \"kubernetes.io/projected/f142fda8-6ffd-4418-b724-d55c78399964-kube-api-access-msr62\") pod \"coredns-66bc5c9577-49znc\" (UID: \"f142fda8-6ffd-4418-b724-d55c78399964\") " pod="kube-system/coredns-66bc5c9577-49znc" Apr 23 23:19:56.912321 kubelet[3427]: I0423 23:19:56.912283 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b94f8cb1-0ece-4e1e-a59c-df0ce7bed609-goldmane-ca-bundle\") pod \"goldmane-6b4b7f4496-j4zb6\" (UID: \"b94f8cb1-0ece-4e1e-a59c-df0ce7bed609\") " pod="calico-system/goldmane-6b4b7f4496-j4zb6" Apr 23 23:19:56.912321 kubelet[3427]: I0423 23:19:56.912307 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b7db\" (UniqueName: \"kubernetes.io/projected/043b430d-28cd-4c76-9d72-4b30392409aa-kube-api-access-8b7db\") pod \"calico-kube-controllers-f94c58f4-b8pwc\" (UID: \"043b430d-28cd-4c76-9d72-4b30392409aa\") " pod="calico-system/calico-kube-controllers-f94c58f4-b8pwc" Apr 23 23:19:56.912321 kubelet[3427]: I0423 23:19:56.912319 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s6gl\" (UniqueName: \"kubernetes.io/projected/9bd25f44-e1d9-4f63-84fb-09a335c9c974-kube-api-access-2s6gl\") pod \"coredns-66bc5c9577-zc5vb\" (UID: \"9bd25f44-e1d9-4f63-84fb-09a335c9c974\") " pod="kube-system/coredns-66bc5c9577-zc5vb" Apr 23 23:19:56.912370 kubelet[3427]: I0423 23:19:56.912329 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/fb79fb64-b4a9-40ad-9ed6-8c74737a8635-calico-apiserver-certs\") pod \"calico-apiserver-57dbd4f578-f7nhj\" (UID: \"fb79fb64-b4a9-40ad-9ed6-8c74737a8635\") " pod="calico-system/calico-apiserver-57dbd4f578-f7nhj" Apr 23 23:19:56.912370 kubelet[3427]: I0423 23:19:56.912339 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/48f3e795-c1d9-4588-9444-c43e72073baf-calico-apiserver-certs\") pod \"calico-apiserver-57dbd4f578-fkj7t\" (UID: \"48f3e795-c1d9-4588-9444-c43e72073baf\") " pod="calico-system/calico-apiserver-57dbd4f578-fkj7t" Apr 23 23:19:56.913434 kubelet[3427]: I0423 23:19:56.912583 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grbzj\" (UniqueName: \"kubernetes.io/projected/48f3e795-c1d9-4588-9444-c43e72073baf-kube-api-access-grbzj\") pod \"calico-apiserver-57dbd4f578-fkj7t\" (UID: \"48f3e795-c1d9-4588-9444-c43e72073baf\") " pod="calico-system/calico-apiserver-57dbd4f578-fkj7t" Apr 23 23:19:56.913434 kubelet[3427]: I0423 23:19:56.912650 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f142fda8-6ffd-4418-b724-d55c78399964-config-volume\") pod \"coredns-66bc5c9577-49znc\" (UID: \"f142fda8-6ffd-4418-b724-d55c78399964\") " pod="kube-system/coredns-66bc5c9577-49znc" Apr 23 23:19:56.913434 kubelet[3427]: I0423 23:19:56.912661 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b94f8cb1-0ece-4e1e-a59c-df0ce7bed609-config\") pod \"goldmane-6b4b7f4496-j4zb6\" (UID: \"b94f8cb1-0ece-4e1e-a59c-df0ce7bed609\") " pod="calico-system/goldmane-6b4b7f4496-j4zb6" Apr 23 23:19:56.913434 kubelet[3427]: I0423 23:19:56.912671 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5bm7\" (UniqueName: \"kubernetes.io/projected/fb79fb64-b4a9-40ad-9ed6-8c74737a8635-kube-api-access-s5bm7\") pod \"calico-apiserver-57dbd4f578-f7nhj\" (UID: \"fb79fb64-b4a9-40ad-9ed6-8c74737a8635\") " pod="calico-system/calico-apiserver-57dbd4f578-f7nhj" Apr 23 23:19:56.913434 kubelet[3427]: I0423 23:19:56.912691 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-ca-bundle\") pod \"whisker-bd885f7c6-5mtnn\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " pod="calico-system/whisker-bd885f7c6-5mtnn" Apr 23 23:19:56.913567 kubelet[3427]: I0423 23:19:56.912712 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/b94f8cb1-0ece-4e1e-a59c-df0ce7bed609-goldmane-key-pair\") pod \"goldmane-6b4b7f4496-j4zb6\" (UID: \"b94f8cb1-0ece-4e1e-a59c-df0ce7bed609\") " pod="calico-system/goldmane-6b4b7f4496-j4zb6" Apr 23 23:19:56.913567 kubelet[3427]: I0423 23:19:56.912722 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7wsl\" (UniqueName: \"kubernetes.io/projected/b94f8cb1-0ece-4e1e-a59c-df0ce7bed609-kube-api-access-r7wsl\") pod \"goldmane-6b4b7f4496-j4zb6\" (UID: \"b94f8cb1-0ece-4e1e-a59c-df0ce7bed609\") " pod="calico-system/goldmane-6b4b7f4496-j4zb6" Apr 23 23:19:57.118343 containerd[1889]: time="2026-04-23T23:19:57.118026017Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f94c58f4-b8pwc,Uid:043b430d-28cd-4c76-9d72-4b30392409aa,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:57.132946 containerd[1889]: time="2026-04-23T23:19:57.132920720Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zc5vb,Uid:9bd25f44-e1d9-4f63-84fb-09a335c9c974,Namespace:kube-system,Attempt:0,}" Apr 23 23:19:57.138492 containerd[1889]: time="2026-04-23T23:19:57.138471409Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-49znc,Uid:f142fda8-6ffd-4418-b724-d55c78399964,Namespace:kube-system,Attempt:0,}" Apr 23 23:19:57.154553 containerd[1889]: time="2026-04-23T23:19:57.154525203Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-6b4b7f4496-j4zb6,Uid:b94f8cb1-0ece-4e1e-a59c-df0ce7bed609,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:57.165479 containerd[1889]: time="2026-04-23T23:19:57.165405956Z" level=error msg="Failed to destroy network for sandbox \"1d5455764e085b54d8be42ba9cd4390cc7c6e79347d36d871ecd2593cf96a771\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.170325 containerd[1889]: time="2026-04-23T23:19:57.170133669Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-f7nhj,Uid:fb79fb64-b4a9-40ad-9ed6-8c74737a8635,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:57.172960 containerd[1889]: time="2026-04-23T23:19:57.172930374Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f94c58f4-b8pwc,Uid:043b430d-28cd-4c76-9d72-4b30392409aa,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d5455764e085b54d8be42ba9cd4390cc7c6e79347d36d871ecd2593cf96a771\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.173912 kubelet[3427]: E0423 23:19:57.173240 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d5455764e085b54d8be42ba9cd4390cc7c6e79347d36d871ecd2593cf96a771\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.173912 kubelet[3427]: E0423 23:19:57.173286 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d5455764e085b54d8be42ba9cd4390cc7c6e79347d36d871ecd2593cf96a771\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-f94c58f4-b8pwc" Apr 23 23:19:57.173912 kubelet[3427]: E0423 23:19:57.173300 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d5455764e085b54d8be42ba9cd4390cc7c6e79347d36d871ecd2593cf96a771\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-f94c58f4-b8pwc" Apr 23 23:19:57.174031 kubelet[3427]: E0423 23:19:57.173335 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-f94c58f4-b8pwc_calico-system(043b430d-28cd-4c76-9d72-4b30392409aa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-f94c58f4-b8pwc_calico-system(043b430d-28cd-4c76-9d72-4b30392409aa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1d5455764e085b54d8be42ba9cd4390cc7c6e79347d36d871ecd2593cf96a771\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-f94c58f4-b8pwc" podUID="043b430d-28cd-4c76-9d72-4b30392409aa" Apr 23 23:19:57.175222 containerd[1889]: time="2026-04-23T23:19:57.175076678Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-bd885f7c6-5mtnn,Uid:9035f2c7-dbfc-4199-a5d8-b57d671b43bb,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:57.184995 containerd[1889]: time="2026-04-23T23:19:57.184865918Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-fkj7t,Uid:48f3e795-c1d9-4588-9444-c43e72073baf,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:57.256265 containerd[1889]: time="2026-04-23T23:19:57.256219572Z" level=error msg="Failed to destroy network for sandbox \"d3bdec1607276a7ad0032cf53de35f91c757eacfc33bda47099c7cc289c678ee\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.259388 containerd[1889]: time="2026-04-23T23:19:57.259346713Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-49znc,Uid:f142fda8-6ffd-4418-b724-d55c78399964,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3bdec1607276a7ad0032cf53de35f91c757eacfc33bda47099c7cc289c678ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.259690 kubelet[3427]: E0423 23:19:57.259553 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3bdec1607276a7ad0032cf53de35f91c757eacfc33bda47099c7cc289c678ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.259690 kubelet[3427]: E0423 23:19:57.259633 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3bdec1607276a7ad0032cf53de35f91c757eacfc33bda47099c7cc289c678ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-49znc" Apr 23 23:19:57.259690 kubelet[3427]: E0423 23:19:57.259649 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3bdec1607276a7ad0032cf53de35f91c757eacfc33bda47099c7cc289c678ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-49znc" Apr 23 23:19:57.259826 kubelet[3427]: E0423 23:19:57.259697 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-49znc_kube-system(f142fda8-6ffd-4418-b724-d55c78399964)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-49znc_kube-system(f142fda8-6ffd-4418-b724-d55c78399964)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d3bdec1607276a7ad0032cf53de35f91c757eacfc33bda47099c7cc289c678ee\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-49znc" podUID="f142fda8-6ffd-4418-b724-d55c78399964" Apr 23 23:19:57.274887 containerd[1889]: time="2026-04-23T23:19:57.274832039Z" level=error msg="Failed to destroy network for sandbox \"d4a005e1caaf6e156255d1c4efde9db84def48da9a78445fa29f7784f0f4125c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.277894 containerd[1889]: time="2026-04-23T23:19:57.277862624Z" level=error msg="Failed to destroy network for sandbox \"45cc1cdd699c43e0c78d1cbd2f90ceaf89111006d95768bf4744bf2660425c94\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.278972 containerd[1889]: time="2026-04-23T23:19:57.278838549Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-6b4b7f4496-j4zb6,Uid:b94f8cb1-0ece-4e1e-a59c-df0ce7bed609,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4a005e1caaf6e156255d1c4efde9db84def48da9a78445fa29f7784f0f4125c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.280766 kubelet[3427]: E0423 23:19:57.280730 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4a005e1caaf6e156255d1c4efde9db84def48da9a78445fa29f7784f0f4125c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.280996 kubelet[3427]: E0423 23:19:57.280970 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4a005e1caaf6e156255d1c4efde9db84def48da9a78445fa29f7784f0f4125c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-6b4b7f4496-j4zb6" Apr 23 23:19:57.280996 kubelet[3427]: E0423 23:19:57.280992 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4a005e1caaf6e156255d1c4efde9db84def48da9a78445fa29f7784f0f4125c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-6b4b7f4496-j4zb6" Apr 23 23:19:57.281057 kubelet[3427]: E0423 23:19:57.281040 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-6b4b7f4496-j4zb6_calico-system(b94f8cb1-0ece-4e1e-a59c-df0ce7bed609)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-6b4b7f4496-j4zb6_calico-system(b94f8cb1-0ece-4e1e-a59c-df0ce7bed609)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d4a005e1caaf6e156255d1c4efde9db84def48da9a78445fa29f7784f0f4125c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-6b4b7f4496-j4zb6" podUID="b94f8cb1-0ece-4e1e-a59c-df0ce7bed609" Apr 23 23:19:57.282788 containerd[1889]: time="2026-04-23T23:19:57.282716214Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zc5vb,Uid:9bd25f44-e1d9-4f63-84fb-09a335c9c974,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"45cc1cdd699c43e0c78d1cbd2f90ceaf89111006d95768bf4744bf2660425c94\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.283035 kubelet[3427]: E0423 23:19:57.282944 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45cc1cdd699c43e0c78d1cbd2f90ceaf89111006d95768bf4744bf2660425c94\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.283035 kubelet[3427]: E0423 23:19:57.282989 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45cc1cdd699c43e0c78d1cbd2f90ceaf89111006d95768bf4744bf2660425c94\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zc5vb" Apr 23 23:19:57.283035 kubelet[3427]: E0423 23:19:57.283002 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45cc1cdd699c43e0c78d1cbd2f90ceaf89111006d95768bf4744bf2660425c94\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zc5vb" Apr 23 23:19:57.283146 kubelet[3427]: E0423 23:19:57.283031 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-zc5vb_kube-system(9bd25f44-e1d9-4f63-84fb-09a335c9c974)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-zc5vb_kube-system(9bd25f44-e1d9-4f63-84fb-09a335c9c974)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"45cc1cdd699c43e0c78d1cbd2f90ceaf89111006d95768bf4744bf2660425c94\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-zc5vb" podUID="9bd25f44-e1d9-4f63-84fb-09a335c9c974" Apr 23 23:19:57.283273 containerd[1889]: time="2026-04-23T23:19:57.283017058Z" level=error msg="Failed to destroy network for sandbox \"103767f24bdac442e928821b0fee28f71751ffd5184e75395e6e7f0fe8b49329\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.286756 containerd[1889]: time="2026-04-23T23:19:57.286727149Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-f7nhj,Uid:fb79fb64-b4a9-40ad-9ed6-8c74737a8635,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"103767f24bdac442e928821b0fee28f71751ffd5184e75395e6e7f0fe8b49329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.287911 kubelet[3427]: E0423 23:19:57.287818 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"103767f24bdac442e928821b0fee28f71751ffd5184e75395e6e7f0fe8b49329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.287911 kubelet[3427]: E0423 23:19:57.287858 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"103767f24bdac442e928821b0fee28f71751ffd5184e75395e6e7f0fe8b49329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57dbd4f578-f7nhj" Apr 23 23:19:57.287911 kubelet[3427]: E0423 23:19:57.287871 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"103767f24bdac442e928821b0fee28f71751ffd5184e75395e6e7f0fe8b49329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57dbd4f578-f7nhj" Apr 23 23:19:57.288018 kubelet[3427]: E0423 23:19:57.287898 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-57dbd4f578-f7nhj_calico-system(fb79fb64-b4a9-40ad-9ed6-8c74737a8635)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-57dbd4f578-f7nhj_calico-system(fb79fb64-b4a9-40ad-9ed6-8c74737a8635)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"103767f24bdac442e928821b0fee28f71751ffd5184e75395e6e7f0fe8b49329\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-57dbd4f578-f7nhj" podUID="fb79fb64-b4a9-40ad-9ed6-8c74737a8635" Apr 23 23:19:57.292471 containerd[1889]: time="2026-04-23T23:19:57.292401626Z" level=error msg="Failed to destroy network for sandbox \"7db8f7533adfecfc9f707323dc99c4df74ac4d701a006b326a039d9fa00d521f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.294358 containerd[1889]: time="2026-04-23T23:19:57.294335114Z" level=error msg="Failed to destroy network for sandbox \"2b9ed8fd8085a2cde3a6a612dc8c1b3a816f76cf61ba3d035f70a639554e9bfd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.296524 containerd[1889]: time="2026-04-23T23:19:57.296491379Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-fkj7t,Uid:48f3e795-c1d9-4588-9444-c43e72073baf,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7db8f7533adfecfc9f707323dc99c4df74ac4d701a006b326a039d9fa00d521f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.297330 kubelet[3427]: E0423 23:19:57.297110 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7db8f7533adfecfc9f707323dc99c4df74ac4d701a006b326a039d9fa00d521f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.297330 kubelet[3427]: E0423 23:19:57.297143 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7db8f7533adfecfc9f707323dc99c4df74ac4d701a006b326a039d9fa00d521f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57dbd4f578-fkj7t" Apr 23 23:19:57.297330 kubelet[3427]: E0423 23:19:57.297156 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7db8f7533adfecfc9f707323dc99c4df74ac4d701a006b326a039d9fa00d521f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57dbd4f578-fkj7t" Apr 23 23:19:57.297465 kubelet[3427]: E0423 23:19:57.297194 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-57dbd4f578-fkj7t_calico-system(48f3e795-c1d9-4588-9444-c43e72073baf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-57dbd4f578-fkj7t_calico-system(48f3e795-c1d9-4588-9444-c43e72073baf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7db8f7533adfecfc9f707323dc99c4df74ac4d701a006b326a039d9fa00d521f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-57dbd4f578-fkj7t" podUID="48f3e795-c1d9-4588-9444-c43e72073baf" Apr 23 23:19:57.300254 containerd[1889]: time="2026-04-23T23:19:57.300213567Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-bd885f7c6-5mtnn,Uid:9035f2c7-dbfc-4199-a5d8-b57d671b43bb,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9ed8fd8085a2cde3a6a612dc8c1b3a816f76cf61ba3d035f70a639554e9bfd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.300744 kubelet[3427]: E0423 23:19:57.300703 3427 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9ed8fd8085a2cde3a6a612dc8c1b3a816f76cf61ba3d035f70a639554e9bfd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 23 23:19:57.300744 kubelet[3427]: E0423 23:19:57.300744 3427 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9ed8fd8085a2cde3a6a612dc8c1b3a816f76cf61ba3d035f70a639554e9bfd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-bd885f7c6-5mtnn" Apr 23 23:19:57.300831 kubelet[3427]: E0423 23:19:57.300763 3427 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9ed8fd8085a2cde3a6a612dc8c1b3a816f76cf61ba3d035f70a639554e9bfd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-bd885f7c6-5mtnn" Apr 23 23:19:57.300831 kubelet[3427]: E0423 23:19:57.300796 3427 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-bd885f7c6-5mtnn_calico-system(9035f2c7-dbfc-4199-a5d8-b57d671b43bb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-bd885f7c6-5mtnn_calico-system(9035f2c7-dbfc-4199-a5d8-b57d671b43bb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2b9ed8fd8085a2cde3a6a612dc8c1b3a816f76cf61ba3d035f70a639554e9bfd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-bd885f7c6-5mtnn" podUID="9035f2c7-dbfc-4199-a5d8-b57d671b43bb" Apr 23 23:19:57.443295 containerd[1889]: time="2026-04-23T23:19:57.442972437Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Apr 23 23:19:57.458699 containerd[1889]: time="2026-04-23T23:19:57.458667098Z" level=info msg="Container a1ffe6edae47b462cc8ad4730313156ed61235b7ade261c491fb76f1af3764e6: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:19:57.476835 containerd[1889]: time="2026-04-23T23:19:57.476546353Z" level=info msg="CreateContainer within sandbox \"efea7e2e8de918bbb7dfd7173ce7e0fde552201f84f5e6f849d5668687982c10\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"a1ffe6edae47b462cc8ad4730313156ed61235b7ade261c491fb76f1af3764e6\"" Apr 23 23:19:57.478187 containerd[1889]: time="2026-04-23T23:19:57.478152790Z" level=info msg="StartContainer for \"a1ffe6edae47b462cc8ad4730313156ed61235b7ade261c491fb76f1af3764e6\"" Apr 23 23:19:57.479630 containerd[1889]: time="2026-04-23T23:19:57.479503528Z" level=info msg="connecting to shim a1ffe6edae47b462cc8ad4730313156ed61235b7ade261c491fb76f1af3764e6" address="unix:///run/containerd/s/53309e2d30f48cf732877c6e74c1f75ec93eb58cc38f7a162cc072a81364c299" protocol=ttrpc version=3 Apr 23 23:19:57.498710 systemd[1]: Started cri-containerd-a1ffe6edae47b462cc8ad4730313156ed61235b7ade261c491fb76f1af3764e6.scope - libcontainer container a1ffe6edae47b462cc8ad4730313156ed61235b7ade261c491fb76f1af3764e6. Apr 23 23:19:57.552346 containerd[1889]: time="2026-04-23T23:19:57.552272851Z" level=info msg="StartContainer for \"a1ffe6edae47b462cc8ad4730313156ed61235b7ade261c491fb76f1af3764e6\" returns successfully" Apr 23 23:19:57.719993 kubelet[3427]: I0423 23:19:57.719891 3427 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvd2d\" (UniqueName: \"kubernetes.io/projected/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-kube-api-access-lvd2d\") pod \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " Apr 23 23:19:57.719993 kubelet[3427]: I0423 23:19:57.719939 3427 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-ca-bundle\") pod \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " Apr 23 23:19:57.719993 kubelet[3427]: I0423 23:19:57.719955 3427 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-nginx-config\") pod \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " Apr 23 23:19:57.719993 kubelet[3427]: I0423 23:19:57.719974 3427 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-backend-key-pair\") pod \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\" (UID: \"9035f2c7-dbfc-4199-a5d8-b57d671b43bb\") " Apr 23 23:19:57.720554 kubelet[3427]: I0423 23:19:57.720403 3427 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "9035f2c7-dbfc-4199-a5d8-b57d671b43bb" (UID: "9035f2c7-dbfc-4199-a5d8-b57d671b43bb"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 23 23:19:57.722146 kubelet[3427]: I0423 23:19:57.721760 3427 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-nginx-config" (OuterVolumeSpecName: "nginx-config") pod "9035f2c7-dbfc-4199-a5d8-b57d671b43bb" (UID: "9035f2c7-dbfc-4199-a5d8-b57d671b43bb"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 23 23:19:57.723630 kubelet[3427]: I0423 23:19:57.723587 3427 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "9035f2c7-dbfc-4199-a5d8-b57d671b43bb" (UID: "9035f2c7-dbfc-4199-a5d8-b57d671b43bb"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 23 23:19:57.724474 kubelet[3427]: I0423 23:19:57.724355 3427 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-kube-api-access-lvd2d" (OuterVolumeSpecName: "kube-api-access-lvd2d") pod "9035f2c7-dbfc-4199-a5d8-b57d671b43bb" (UID: "9035f2c7-dbfc-4199-a5d8-b57d671b43bb"). InnerVolumeSpecName "kube-api-access-lvd2d". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 23 23:19:57.820503 kubelet[3427]: I0423 23:19:57.820478 3427 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-backend-key-pair\") on node \"ci-4459.2.4-n-8b75ae451f\" DevicePath \"\"" Apr 23 23:19:57.820503 kubelet[3427]: I0423 23:19:57.820501 3427 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lvd2d\" (UniqueName: \"kubernetes.io/projected/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-kube-api-access-lvd2d\") on node \"ci-4459.2.4-n-8b75ae451f\" DevicePath \"\"" Apr 23 23:19:57.820611 kubelet[3427]: I0423 23:19:57.820513 3427 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-whisker-ca-bundle\") on node \"ci-4459.2.4-n-8b75ae451f\" DevicePath \"\"" Apr 23 23:19:57.820611 kubelet[3427]: I0423 23:19:57.820520 3427 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/9035f2c7-dbfc-4199-a5d8-b57d671b43bb-nginx-config\") on node \"ci-4459.2.4-n-8b75ae451f\" DevicePath \"\"" Apr 23 23:19:58.022660 systemd[1]: var-lib-kubelet-pods-9035f2c7\x2ddbfc\x2d4199\x2da5d8\x2db57d671b43bb-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Apr 23 23:19:58.325108 systemd[1]: Removed slice kubepods-besteffort-pod9035f2c7_dbfc_4199_a5d8_b57d671b43bb.slice - libcontainer container kubepods-besteffort-pod9035f2c7_dbfc_4199_a5d8_b57d671b43bb.slice. Apr 23 23:19:58.452896 kubelet[3427]: I0423 23:19:58.452739 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-zzgnk" podStartSLOduration=4.898658736 podStartE2EDuration="17.452724462s" podCreationTimestamp="2026-04-23 23:19:41 +0000 UTC" firstStartedPulling="2026-04-23 23:19:42.069128912 +0000 UTC m=+15.821250943" lastFinishedPulling="2026-04-23 23:19:54.62319463 +0000 UTC m=+28.375316669" observedRunningTime="2026-04-23 23:19:58.452066198 +0000 UTC m=+32.204188277" watchObservedRunningTime="2026-04-23 23:19:58.452724462 +0000 UTC m=+32.204846493" Apr 23 23:19:58.524431 systemd[1]: Created slice kubepods-besteffort-pod6d0120cd_349a_425a_9f48_1f46a6e7749f.slice - libcontainer container kubepods-besteffort-pod6d0120cd_349a_425a_9f48_1f46a6e7749f.slice. Apr 23 23:19:58.624036 kubelet[3427]: I0423 23:19:58.623933 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/6d0120cd-349a-425a-9f48-1f46a6e7749f-whisker-backend-key-pair\") pod \"whisker-6858f474b6-m597s\" (UID: \"6d0120cd-349a-425a-9f48-1f46a6e7749f\") " pod="calico-system/whisker-6858f474b6-m597s" Apr 23 23:19:58.624036 kubelet[3427]: I0423 23:19:58.623971 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d0120cd-349a-425a-9f48-1f46a6e7749f-whisker-ca-bundle\") pod \"whisker-6858f474b6-m597s\" (UID: \"6d0120cd-349a-425a-9f48-1f46a6e7749f\") " pod="calico-system/whisker-6858f474b6-m597s" Apr 23 23:19:58.624036 kubelet[3427]: I0423 23:19:58.623991 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/6d0120cd-349a-425a-9f48-1f46a6e7749f-nginx-config\") pod \"whisker-6858f474b6-m597s\" (UID: \"6d0120cd-349a-425a-9f48-1f46a6e7749f\") " pod="calico-system/whisker-6858f474b6-m597s" Apr 23 23:19:58.624036 kubelet[3427]: I0423 23:19:58.624001 3427 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2nz2\" (UniqueName: \"kubernetes.io/projected/6d0120cd-349a-425a-9f48-1f46a6e7749f-kube-api-access-r2nz2\") pod \"whisker-6858f474b6-m597s\" (UID: \"6d0120cd-349a-425a-9f48-1f46a6e7749f\") " pod="calico-system/whisker-6858f474b6-m597s" Apr 23 23:19:58.833632 containerd[1889]: time="2026-04-23T23:19:58.833424239Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6858f474b6-m597s,Uid:6d0120cd-349a-425a-9f48-1f46a6e7749f,Namespace:calico-system,Attempt:0,}" Apr 23 23:19:58.966085 systemd-networkd[1474]: cali82c3cde235b: Link UP Apr 23 23:19:58.966195 systemd-networkd[1474]: cali82c3cde235b: Gained carrier Apr 23 23:19:58.982970 containerd[1889]: 2026-04-23 23:19:58.859 [ERROR][4532] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 23 23:19:58.982970 containerd[1889]: 2026-04-23 23:19:58.870 [INFO][4532] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0 whisker-6858f474b6- calico-system 6d0120cd-349a-425a-9f48-1f46a6e7749f 869 0 2026-04-23 23:19:58 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:6858f474b6 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f whisker-6858f474b6-m597s eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali82c3cde235b [] [] }} ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-" Apr 23 23:19:58.982970 containerd[1889]: 2026-04-23 23:19:58.870 [INFO][4532] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" Apr 23 23:19:58.982970 containerd[1889]: 2026-04-23 23:19:58.898 [INFO][4570] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" HandleID="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Workload="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.906 [INFO][4570] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" HandleID="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Workload="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002efdd0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"whisker-6858f474b6-m597s", "timestamp":"2026-04-23 23:19:58.898049439 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000405b80)} Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.906 [INFO][4570] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.906 [INFO][4570] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.906 [INFO][4570] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.910 [INFO][4570] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.915 [INFO][4570] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.919 [INFO][4570] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.921 [INFO][4570] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983437 containerd[1889]: 2026-04-23 23:19:58.922 [INFO][4570] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983991 containerd[1889]: 2026-04-23 23:19:58.923 [INFO][4570] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983991 containerd[1889]: 2026-04-23 23:19:58.924 [INFO][4570] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda Apr 23 23:19:58.983991 containerd[1889]: 2026-04-23 23:19:58.928 [INFO][4570] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983991 containerd[1889]: 2026-04-23 23:19:58.937 [INFO][4570] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.129/26] block=192.168.24.128/26 handle="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983991 containerd[1889]: 2026-04-23 23:19:58.937 [INFO][4570] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.129/26] handle="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:19:58.983991 containerd[1889]: 2026-04-23 23:19:58.937 [INFO][4570] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:19:58.983991 containerd[1889]: 2026-04-23 23:19:58.937 [INFO][4570] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.129/26] IPv6=[] ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" HandleID="k8s-pod-network.6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Workload="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" Apr 23 23:19:58.984686 containerd[1889]: 2026-04-23 23:19:58.940 [INFO][4532] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0", GenerateName:"whisker-6858f474b6-", Namespace:"calico-system", SelfLink:"", UID:"6d0120cd-349a-425a-9f48-1f46a6e7749f", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"6858f474b6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"whisker-6858f474b6-m597s", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.24.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali82c3cde235b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:19:58.984686 containerd[1889]: 2026-04-23 23:19:58.940 [INFO][4532] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.129/32] ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" Apr 23 23:19:58.984860 containerd[1889]: 2026-04-23 23:19:58.940 [INFO][4532] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali82c3cde235b ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" Apr 23 23:19:58.984860 containerd[1889]: 2026-04-23 23:19:58.965 [INFO][4532] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" Apr 23 23:19:58.984901 containerd[1889]: 2026-04-23 23:19:58.965 [INFO][4532] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0", GenerateName:"whisker-6858f474b6-", Namespace:"calico-system", SelfLink:"", UID:"6d0120cd-349a-425a-9f48-1f46a6e7749f", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"6858f474b6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda", Pod:"whisker-6858f474b6-m597s", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.24.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali82c3cde235b", MAC:"e2:39:18:cc:93:33", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:19:58.985302 containerd[1889]: 2026-04-23 23:19:58.979 [INFO][4532] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" Namespace="calico-system" Pod="whisker-6858f474b6-m597s" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-whisker--6858f474b6--m597s-eth0" Apr 23 23:19:59.043120 containerd[1889]: time="2026-04-23T23:19:59.043023124Z" level=info msg="connecting to shim 6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda" address="unix:///run/containerd/s/18b355e030c60f370bf0d24cf925b88a62e2e71465b952cf8647fea97f2c39ef" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:19:59.081070 systemd[1]: Started cri-containerd-6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda.scope - libcontainer container 6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda. Apr 23 23:19:59.115940 containerd[1889]: time="2026-04-23T23:19:59.115909528Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6858f474b6-m597s,Uid:6d0120cd-349a-425a-9f48-1f46a6e7749f,Namespace:calico-system,Attempt:0,} returns sandbox id \"6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda\"" Apr 23 23:19:59.117465 containerd[1889]: time="2026-04-23T23:19:59.117440344Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.5\"" Apr 23 23:20:00.328288 kubelet[3427]: I0423 23:20:00.328080 3427 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9035f2c7-dbfc-4199-a5d8-b57d671b43bb" path="/var/lib/kubelet/pods/9035f2c7-dbfc-4199-a5d8-b57d671b43bb/volumes" Apr 23 23:20:00.603921 containerd[1889]: time="2026-04-23T23:20:00.603813949Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:00.606617 containerd[1889]: time="2026-04-23T23:20:00.606566225Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.5: active requests=0, bytes read=5896864" Apr 23 23:20:00.610075 containerd[1889]: time="2026-04-23T23:20:00.609943012Z" level=info msg="ImageCreate event name:\"sha256:a47d4844a7d3a4350ed0ac1bc7a5e68be5c0d8a9b81906debd805ec9c4deec82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:00.615499 containerd[1889]: time="2026-04-23T23:20:00.615118505Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:b143cf26c347546feabb95cec04a2349f5ae297830cc54fdc2578b89d1a3e021\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:00.616119 containerd[1889]: time="2026-04-23T23:20:00.616083996Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.5\" with image id \"sha256:a47d4844a7d3a4350ed0ac1bc7a5e68be5c0d8a9b81906debd805ec9c4deec82\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:b143cf26c347546feabb95cec04a2349f5ae297830cc54fdc2578b89d1a3e021\", size \"8472495\" in 1.498613467s" Apr 23 23:20:00.616119 containerd[1889]: time="2026-04-23T23:20:00.616113509Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.5\" returns image reference \"sha256:a47d4844a7d3a4350ed0ac1bc7a5e68be5c0d8a9b81906debd805ec9c4deec82\"" Apr 23 23:20:00.625066 containerd[1889]: time="2026-04-23T23:20:00.625022161Z" level=info msg="CreateContainer within sandbox \"6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Apr 23 23:20:00.647525 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2347038302.mount: Deactivated successfully. Apr 23 23:20:00.647768 containerd[1889]: time="2026-04-23T23:20:00.647682890Z" level=info msg="Container b987598ab06d1b272b3837d91c7c05d796ebb93f64609f2c10dd088a39ccabfa: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:00.666995 containerd[1889]: time="2026-04-23T23:20:00.666949879Z" level=info msg="CreateContainer within sandbox \"6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"b987598ab06d1b272b3837d91c7c05d796ebb93f64609f2c10dd088a39ccabfa\"" Apr 23 23:20:00.668025 containerd[1889]: time="2026-04-23T23:20:00.667387087Z" level=info msg="StartContainer for \"b987598ab06d1b272b3837d91c7c05d796ebb93f64609f2c10dd088a39ccabfa\"" Apr 23 23:20:00.668918 containerd[1889]: time="2026-04-23T23:20:00.668877357Z" level=info msg="connecting to shim b987598ab06d1b272b3837d91c7c05d796ebb93f64609f2c10dd088a39ccabfa" address="unix:///run/containerd/s/18b355e030c60f370bf0d24cf925b88a62e2e71465b952cf8647fea97f2c39ef" protocol=ttrpc version=3 Apr 23 23:20:00.686713 systemd[1]: Started cri-containerd-b987598ab06d1b272b3837d91c7c05d796ebb93f64609f2c10dd088a39ccabfa.scope - libcontainer container b987598ab06d1b272b3837d91c7c05d796ebb93f64609f2c10dd088a39ccabfa. Apr 23 23:20:00.720871 containerd[1889]: time="2026-04-23T23:20:00.720804752Z" level=info msg="StartContainer for \"b987598ab06d1b272b3837d91c7c05d796ebb93f64609f2c10dd088a39ccabfa\" returns successfully" Apr 23 23:20:00.722034 containerd[1889]: time="2026-04-23T23:20:00.722009483Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\"" Apr 23 23:20:00.732724 systemd-networkd[1474]: vxlan.calico: Link UP Apr 23 23:20:00.732729 systemd-networkd[1474]: vxlan.calico: Gained carrier Apr 23 23:20:00.749475 systemd-networkd[1474]: cali82c3cde235b: Gained IPv6LL Apr 23 23:20:02.446896 kubelet[3427]: I0423 23:20:02.446122 3427 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 23 23:20:02.476998 systemd-networkd[1474]: vxlan.calico: Gained IPv6LL Apr 23 23:20:02.662408 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1656350045.mount: Deactivated successfully. Apr 23 23:20:02.716956 containerd[1889]: time="2026-04-23T23:20:02.716401210Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:02.719259 containerd[1889]: time="2026-04-23T23:20:02.719237449Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.5: active requests=0, bytes read=15624823" Apr 23 23:20:02.722470 containerd[1889]: time="2026-04-23T23:20:02.722438702Z" level=info msg="ImageCreate event name:\"sha256:b6ad9a1ad05ff3a8548f5adf860703add7bc41ef2f24f47e461f1914f73f7c8f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:02.733312 containerd[1889]: time="2026-04-23T23:20:02.733261216Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:0bec142ebaa70bcdda5553c7316abcef9cb60a35c2e3ed16b75f26313de91eed\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:02.733866 containerd[1889]: time="2026-04-23T23:20:02.733841509Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\" with image id \"sha256:b6ad9a1ad05ff3a8548f5adf860703add7bc41ef2f24f47e461f1914f73f7c8f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:0bec142ebaa70bcdda5553c7316abcef9cb60a35c2e3ed16b75f26313de91eed\", size \"15624653\" in 2.011804505s" Apr 23 23:20:02.733962 containerd[1889]: time="2026-04-23T23:20:02.733942537Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\" returns image reference \"sha256:b6ad9a1ad05ff3a8548f5adf860703add7bc41ef2f24f47e461f1914f73f7c8f\"" Apr 23 23:20:02.741023 containerd[1889]: time="2026-04-23T23:20:02.740627172Z" level=info msg="CreateContainer within sandbox \"6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Apr 23 23:20:02.763745 containerd[1889]: time="2026-04-23T23:20:02.760691814Z" level=info msg="Container 48d4247ffcee2f2f6f453a575bad7a4854677f1803f372a3f90979fc1ffda916: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:02.780100 containerd[1889]: time="2026-04-23T23:20:02.780064743Z" level=info msg="CreateContainer within sandbox \"6e1ac32262357bb2c759d157a994bfb6ccfc32e0d64bca460d83858dda29dfda\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"48d4247ffcee2f2f6f453a575bad7a4854677f1803f372a3f90979fc1ffda916\"" Apr 23 23:20:02.780622 containerd[1889]: time="2026-04-23T23:20:02.780581722Z" level=info msg="StartContainer for \"48d4247ffcee2f2f6f453a575bad7a4854677f1803f372a3f90979fc1ffda916\"" Apr 23 23:20:02.781655 containerd[1889]: time="2026-04-23T23:20:02.781628184Z" level=info msg="connecting to shim 48d4247ffcee2f2f6f453a575bad7a4854677f1803f372a3f90979fc1ffda916" address="unix:///run/containerd/s/18b355e030c60f370bf0d24cf925b88a62e2e71465b952cf8647fea97f2c39ef" protocol=ttrpc version=3 Apr 23 23:20:02.802709 systemd[1]: Started cri-containerd-48d4247ffcee2f2f6f453a575bad7a4854677f1803f372a3f90979fc1ffda916.scope - libcontainer container 48d4247ffcee2f2f6f453a575bad7a4854677f1803f372a3f90979fc1ffda916. Apr 23 23:20:02.837298 containerd[1889]: time="2026-04-23T23:20:02.837271954Z" level=info msg="StartContainer for \"48d4247ffcee2f2f6f453a575bad7a4854677f1803f372a3f90979fc1ffda916\" returns successfully" Apr 23 23:20:03.459882 kubelet[3427]: I0423 23:20:03.459771 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-6858f474b6-m597s" podStartSLOduration=1.842136504 podStartE2EDuration="5.459755617s" podCreationTimestamp="2026-04-23 23:19:58 +0000 UTC" firstStartedPulling="2026-04-23 23:19:59.1171121 +0000 UTC m=+32.869234131" lastFinishedPulling="2026-04-23 23:20:02.734731213 +0000 UTC m=+36.486853244" observedRunningTime="2026-04-23 23:20:03.459005126 +0000 UTC m=+37.211127157" watchObservedRunningTime="2026-04-23 23:20:03.459755617 +0000 UTC m=+37.211877648" Apr 23 23:20:08.327268 containerd[1889]: time="2026-04-23T23:20:08.326977982Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-49znc,Uid:f142fda8-6ffd-4418-b724-d55c78399964,Namespace:kube-system,Attempt:0,}" Apr 23 23:20:08.334565 containerd[1889]: time="2026-04-23T23:20:08.334532185Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f94c58f4-b8pwc,Uid:043b430d-28cd-4c76-9d72-4b30392409aa,Namespace:calico-system,Attempt:0,}" Apr 23 23:20:08.339536 containerd[1889]: time="2026-04-23T23:20:08.339500510Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mlsd9,Uid:290a2228-90ec-4821-be49-5abe8a0f00b0,Namespace:calico-system,Attempt:0,}" Apr 23 23:20:08.478273 systemd-networkd[1474]: cali623229afbfa: Link UP Apr 23 23:20:08.479515 systemd-networkd[1474]: cali623229afbfa: Gained carrier Apr 23 23:20:08.495220 containerd[1889]: 2026-04-23 23:20:08.393 [INFO][4943] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0 coredns-66bc5c9577- kube-system f142fda8-6ffd-4418-b724-d55c78399964 812 0 2026-04-23 23:19:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f coredns-66bc5c9577-49znc eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali623229afbfa [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-" Apr 23 23:20:08.495220 containerd[1889]: 2026-04-23 23:20:08.393 [INFO][4943] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" Apr 23 23:20:08.495220 containerd[1889]: 2026-04-23 23:20:08.423 [INFO][4981] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" HandleID="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Workload="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.432 [INFO][4981] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" HandleID="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Workload="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003776b0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"coredns-66bc5c9577-49znc", "timestamp":"2026-04-23 23:20:08.423762738 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400040edc0)} Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.432 [INFO][4981] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.432 [INFO][4981] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.433 [INFO][4981] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.436 [INFO][4981] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.442 [INFO][4981] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.454 [INFO][4981] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.456 [INFO][4981] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495417 containerd[1889]: 2026-04-23 23:20:08.458 [INFO][4981] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495555 containerd[1889]: 2026-04-23 23:20:08.458 [INFO][4981] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495555 containerd[1889]: 2026-04-23 23:20:08.459 [INFO][4981] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9 Apr 23 23:20:08.495555 containerd[1889]: 2026-04-23 23:20:08.464 [INFO][4981] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495555 containerd[1889]: 2026-04-23 23:20:08.471 [INFO][4981] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.130/26] block=192.168.24.128/26 handle="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495555 containerd[1889]: 2026-04-23 23:20:08.471 [INFO][4981] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.130/26] handle="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.495555 containerd[1889]: 2026-04-23 23:20:08.472 [INFO][4981] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:20:08.495555 containerd[1889]: 2026-04-23 23:20:08.472 [INFO][4981] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.130/26] IPv6=[] ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" HandleID="k8s-pod-network.7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Workload="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" Apr 23 23:20:08.495831 containerd[1889]: 2026-04-23 23:20:08.474 [INFO][4943] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"f142fda8-6ffd-4418-b724-d55c78399964", ResourceVersion:"812", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"coredns-66bc5c9577-49znc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.24.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali623229afbfa", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:08.495831 containerd[1889]: 2026-04-23 23:20:08.474 [INFO][4943] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.130/32] ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" Apr 23 23:20:08.495831 containerd[1889]: 2026-04-23 23:20:08.474 [INFO][4943] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali623229afbfa ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" Apr 23 23:20:08.495831 containerd[1889]: 2026-04-23 23:20:08.479 [INFO][4943] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" Apr 23 23:20:08.495831 containerd[1889]: 2026-04-23 23:20:08.479 [INFO][4943] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"f142fda8-6ffd-4418-b724-d55c78399964", ResourceVersion:"812", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9", Pod:"coredns-66bc5c9577-49znc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.24.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali623229afbfa", MAC:"e6:88:f4:c9:e9:fb", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:08.496056 containerd[1889]: 2026-04-23 23:20:08.492 [INFO][4943] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" Namespace="kube-system" Pod="coredns-66bc5c9577-49znc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--49znc-eth0" Apr 23 23:20:08.554335 containerd[1889]: time="2026-04-23T23:20:08.554267558Z" level=info msg="connecting to shim 7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9" address="unix:///run/containerd/s/7fd8bc2213b8f4b6ac54e2d6715ef2eb2228157593b9ab8a13b8e784e01f9c98" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:20:08.578054 systemd[1]: Started cri-containerd-7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9.scope - libcontainer container 7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9. Apr 23 23:20:08.581282 systemd-networkd[1474]: calie8f5711a37c: Link UP Apr 23 23:20:08.581386 systemd-networkd[1474]: calie8f5711a37c: Gained carrier Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.394 [INFO][4947] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0 calico-kube-controllers-f94c58f4- calico-system 043b430d-28cd-4c76-9d72-4b30392409aa 810 0 2026-04-23 23:19:41 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:f94c58f4 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f calico-kube-controllers-f94c58f4-b8pwc eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calie8f5711a37c [] [] }} ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.394 [INFO][4947] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.438 [INFO][4979] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" HandleID="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.451 [INFO][4979] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" HandleID="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002efdd0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"calico-kube-controllers-f94c58f4-b8pwc", "timestamp":"2026-04-23 23:20:08.438285401 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000375b80)} Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.451 [INFO][4979] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.471 [INFO][4979] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.471 [INFO][4979] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.537 [INFO][4979] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.541 [INFO][4979] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.551 [INFO][4979] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.553 [INFO][4979] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.556 [INFO][4979] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.556 [INFO][4979] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.557 [INFO][4979] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442 Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.566 [INFO][4979] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.572 [INFO][4979] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.131/26] block=192.168.24.128/26 handle="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.572 [INFO][4979] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.131/26] handle="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.573 [INFO][4979] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:20:08.600444 containerd[1889]: 2026-04-23 23:20:08.573 [INFO][4979] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.131/26] IPv6=[] ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" HandleID="k8s-pod-network.c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" Apr 23 23:20:08.601469 containerd[1889]: 2026-04-23 23:20:08.577 [INFO][4947] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0", GenerateName:"calico-kube-controllers-f94c58f4-", Namespace:"calico-system", SelfLink:"", UID:"043b430d-28cd-4c76-9d72-4b30392409aa", ResourceVersion:"810", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f94c58f4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"calico-kube-controllers-f94c58f4-b8pwc", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.24.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie8f5711a37c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:08.601469 containerd[1889]: 2026-04-23 23:20:08.577 [INFO][4947] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.131/32] ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" Apr 23 23:20:08.601469 containerd[1889]: 2026-04-23 23:20:08.577 [INFO][4947] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie8f5711a37c ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" Apr 23 23:20:08.601469 containerd[1889]: 2026-04-23 23:20:08.580 [INFO][4947] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" Apr 23 23:20:08.601469 containerd[1889]: 2026-04-23 23:20:08.582 [INFO][4947] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0", GenerateName:"calico-kube-controllers-f94c58f4-", Namespace:"calico-system", SelfLink:"", UID:"043b430d-28cd-4c76-9d72-4b30392409aa", ResourceVersion:"810", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f94c58f4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442", Pod:"calico-kube-controllers-f94c58f4-b8pwc", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.24.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie8f5711a37c", MAC:"96:2a:03:74:79:94", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:08.601469 containerd[1889]: 2026-04-23 23:20:08.598 [INFO][4947] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" Namespace="calico-system" Pod="calico-kube-controllers-f94c58f4-b8pwc" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--kube--controllers--f94c58f4--b8pwc-eth0" Apr 23 23:20:08.625863 containerd[1889]: time="2026-04-23T23:20:08.625809346Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-49znc,Uid:f142fda8-6ffd-4418-b724-d55c78399964,Namespace:kube-system,Attempt:0,} returns sandbox id \"7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9\"" Apr 23 23:20:08.635698 containerd[1889]: time="2026-04-23T23:20:08.635656024Z" level=info msg="CreateContainer within sandbox \"7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 23 23:20:08.647764 containerd[1889]: time="2026-04-23T23:20:08.647708990Z" level=info msg="connecting to shim c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442" address="unix:///run/containerd/s/56a17a9806fdbcf079b57de62b92931c0df50105ee9da5ad521e5d1eff134ae2" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:20:08.664495 containerd[1889]: time="2026-04-23T23:20:08.664464670Z" level=info msg="Container ab97f86b75eaa34653fad76641ba196dba4ead487ddbca819ac353f4a280b16b: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:08.673719 systemd[1]: Started cri-containerd-c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442.scope - libcontainer container c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442. Apr 23 23:20:08.683340 systemd-networkd[1474]: calia07ea4d2caa: Link UP Apr 23 23:20:08.683954 systemd-networkd[1474]: calia07ea4d2caa: Gained carrier Apr 23 23:20:08.685304 containerd[1889]: time="2026-04-23T23:20:08.683995012Z" level=info msg="CreateContainer within sandbox \"7201900f92f118c255904d2ea6b25906c42f904df3ed814f32deb846c4443ec9\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ab97f86b75eaa34653fad76641ba196dba4ead487ddbca819ac353f4a280b16b\"" Apr 23 23:20:08.688629 containerd[1889]: time="2026-04-23T23:20:08.688072638Z" level=info msg="StartContainer for \"ab97f86b75eaa34653fad76641ba196dba4ead487ddbca819ac353f4a280b16b\"" Apr 23 23:20:08.692761 containerd[1889]: time="2026-04-23T23:20:08.692737350Z" level=info msg="connecting to shim ab97f86b75eaa34653fad76641ba196dba4ead487ddbca819ac353f4a280b16b" address="unix:///run/containerd/s/7fd8bc2213b8f4b6ac54e2d6715ef2eb2228157593b9ab8a13b8e784e01f9c98" protocol=ttrpc version=3 Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.408 [INFO][4965] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0 csi-node-driver- calico-system 290a2228-90ec-4821-be49-5abe8a0f00b0 677 0 2026-04-23 23:19:41 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:95f96f7df k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f csi-node-driver-mlsd9 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calia07ea4d2caa [] [] }} ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.409 [INFO][4965] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.448 [INFO][4990] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" HandleID="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Workload="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.455 [INFO][4990] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" HandleID="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Workload="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002efe70), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"csi-node-driver-mlsd9", "timestamp":"2026-04-23 23:20:08.448009979 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400038b8c0)} Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.455 [INFO][4990] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.573 [INFO][4990] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.575 [INFO][4990] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.637 [INFO][4990] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.642 [INFO][4990] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.651 [INFO][4990] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.654 [INFO][4990] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.656 [INFO][4990] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.658 [INFO][4990] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.660 [INFO][4990] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.668 [INFO][4990] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.676 [INFO][4990] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.132/26] block=192.168.24.128/26 handle="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.676 [INFO][4990] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.132/26] handle="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.676 [INFO][4990] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:20:08.709922 containerd[1889]: 2026-04-23 23:20:08.676 [INFO][4990] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.132/26] IPv6=[] ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" HandleID="k8s-pod-network.1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Workload="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" Apr 23 23:20:08.710609 containerd[1889]: 2026-04-23 23:20:08.678 [INFO][4965] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"290a2228-90ec-4821-be49-5abe8a0f00b0", ResourceVersion:"677", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"95f96f7df", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"csi-node-driver-mlsd9", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.24.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia07ea4d2caa", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:08.710609 containerd[1889]: 2026-04-23 23:20:08.679 [INFO][4965] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.132/32] ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" Apr 23 23:20:08.710609 containerd[1889]: 2026-04-23 23:20:08.679 [INFO][4965] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia07ea4d2caa ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" Apr 23 23:20:08.710609 containerd[1889]: 2026-04-23 23:20:08.684 [INFO][4965] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" Apr 23 23:20:08.710609 containerd[1889]: 2026-04-23 23:20:08.686 [INFO][4965] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"290a2228-90ec-4821-be49-5abe8a0f00b0", ResourceVersion:"677", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"95f96f7df", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d", Pod:"csi-node-driver-mlsd9", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.24.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia07ea4d2caa", MAC:"4a:fd:43:eb:66:ee", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:08.710609 containerd[1889]: 2026-04-23 23:20:08.701 [INFO][4965] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" Namespace="calico-system" Pod="csi-node-driver-mlsd9" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-csi--node--driver--mlsd9-eth0" Apr 23 23:20:08.718799 systemd[1]: Started cri-containerd-ab97f86b75eaa34653fad76641ba196dba4ead487ddbca819ac353f4a280b16b.scope - libcontainer container ab97f86b75eaa34653fad76641ba196dba4ead487ddbca819ac353f4a280b16b. Apr 23 23:20:08.738051 containerd[1889]: time="2026-04-23T23:20:08.737968287Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f94c58f4-b8pwc,Uid:043b430d-28cd-4c76-9d72-4b30392409aa,Namespace:calico-system,Attempt:0,} returns sandbox id \"c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442\"" Apr 23 23:20:08.739692 containerd[1889]: time="2026-04-23T23:20:08.739662242Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\"" Apr 23 23:20:08.756771 containerd[1889]: time="2026-04-23T23:20:08.756750000Z" level=info msg="StartContainer for \"ab97f86b75eaa34653fad76641ba196dba4ead487ddbca819ac353f4a280b16b\" returns successfully" Apr 23 23:20:08.767860 containerd[1889]: time="2026-04-23T23:20:08.767828919Z" level=info msg="connecting to shim 1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d" address="unix:///run/containerd/s/0c3e76de287c9643562696b5d38690f55425f44dc995173038811e32386e5baa" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:20:08.791758 systemd[1]: Started cri-containerd-1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d.scope - libcontainer container 1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d. Apr 23 23:20:08.814325 containerd[1889]: time="2026-04-23T23:20:08.814282968Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mlsd9,Uid:290a2228-90ec-4821-be49-5abe8a0f00b0,Namespace:calico-system,Attempt:0,} returns sandbox id \"1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d\"" Apr 23 23:20:09.322617 containerd[1889]: time="2026-04-23T23:20:09.322531168Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-f7nhj,Uid:fb79fb64-b4a9-40ad-9ed6-8c74737a8635,Namespace:calico-system,Attempt:0,}" Apr 23 23:20:09.457069 systemd-networkd[1474]: cali78b03fbf1af: Link UP Apr 23 23:20:09.457680 systemd-networkd[1474]: cali78b03fbf1af: Gained carrier Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.359 [INFO][5202] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0 calico-apiserver-57dbd4f578- calico-system fb79fb64-b4a9-40ad-9ed6-8c74737a8635 815 0 2026-04-23 23:19:40 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:57dbd4f578 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f calico-apiserver-57dbd4f578-f7nhj eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali78b03fbf1af [] [] }} ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.359 [INFO][5202] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.387 [INFO][5213] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" HandleID="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.398 [INFO][5213] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" HandleID="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ff630), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"calico-apiserver-57dbd4f578-f7nhj", "timestamp":"2026-04-23 23:20:09.387883982 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000208b00)} Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.398 [INFO][5213] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.398 [INFO][5213] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.398 [INFO][5213] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.402 [INFO][5213] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.406 [INFO][5213] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.413 [INFO][5213] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.415 [INFO][5213] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.418 [INFO][5213] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.419 [INFO][5213] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.428 [INFO][5213] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3 Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.434 [INFO][5213] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.451 [INFO][5213] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.133/26] block=192.168.24.128/26 handle="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.451 [INFO][5213] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.133/26] handle="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.451 [INFO][5213] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:20:09.476588 containerd[1889]: 2026-04-23 23:20:09.451 [INFO][5213] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.133/26] IPv6=[] ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" HandleID="k8s-pod-network.ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" Apr 23 23:20:09.479728 containerd[1889]: 2026-04-23 23:20:09.453 [INFO][5202] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0", GenerateName:"calico-apiserver-57dbd4f578-", Namespace:"calico-system", SelfLink:"", UID:"fb79fb64-b4a9-40ad-9ed6-8c74737a8635", ResourceVersion:"815", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57dbd4f578", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"calico-apiserver-57dbd4f578-f7nhj", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.24.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali78b03fbf1af", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:09.479728 containerd[1889]: 2026-04-23 23:20:09.453 [INFO][5202] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.133/32] ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" Apr 23 23:20:09.479728 containerd[1889]: 2026-04-23 23:20:09.453 [INFO][5202] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali78b03fbf1af ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" Apr 23 23:20:09.479728 containerd[1889]: 2026-04-23 23:20:09.459 [INFO][5202] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" Apr 23 23:20:09.479728 containerd[1889]: 2026-04-23 23:20:09.460 [INFO][5202] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0", GenerateName:"calico-apiserver-57dbd4f578-", Namespace:"calico-system", SelfLink:"", UID:"fb79fb64-b4a9-40ad-9ed6-8c74737a8635", ResourceVersion:"815", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57dbd4f578", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3", Pod:"calico-apiserver-57dbd4f578-f7nhj", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.24.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali78b03fbf1af", MAC:"96:fd:0a:95:71:4a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:09.479728 containerd[1889]: 2026-04-23 23:20:09.474 [INFO][5202] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-f7nhj" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--f7nhj-eth0" Apr 23 23:20:09.505564 kubelet[3427]: I0423 23:20:09.504960 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-49znc" podStartSLOduration=38.504946462 podStartE2EDuration="38.504946462s" podCreationTimestamp="2026-04-23 23:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:20:09.488922323 +0000 UTC m=+43.241044378" watchObservedRunningTime="2026-04-23 23:20:09.504946462 +0000 UTC m=+43.257068493" Apr 23 23:20:09.522866 containerd[1889]: time="2026-04-23T23:20:09.522832235Z" level=info msg="connecting to shim ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3" address="unix:///run/containerd/s/2fdf6c93fff9888e73586dfd72eafdd8516412a545e8f62716783c6dad686ae9" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:20:09.543740 systemd[1]: Started cri-containerd-ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3.scope - libcontainer container ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3. Apr 23 23:20:09.572830 containerd[1889]: time="2026-04-23T23:20:09.572731109Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-f7nhj,Uid:fb79fb64-b4a9-40ad-9ed6-8c74737a8635,Namespace:calico-system,Attempt:0,} returns sandbox id \"ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3\"" Apr 23 23:20:09.708682 systemd-networkd[1474]: calia07ea4d2caa: Gained IPv6LL Apr 23 23:20:09.836949 systemd-networkd[1474]: cali623229afbfa: Gained IPv6LL Apr 23 23:20:10.324426 containerd[1889]: time="2026-04-23T23:20:10.324327114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-6b4b7f4496-j4zb6,Uid:b94f8cb1-0ece-4e1e-a59c-df0ce7bed609,Namespace:calico-system,Attempt:0,}" Apr 23 23:20:10.429555 systemd-networkd[1474]: cali7373238b47e: Link UP Apr 23 23:20:10.430415 systemd-networkd[1474]: cali7373238b47e: Gained carrier Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.357 [INFO][5276] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0 goldmane-6b4b7f4496- calico-system b94f8cb1-0ece-4e1e-a59c-df0ce7bed609 813 0 2026-04-23 23:19:41 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:6b4b7f4496 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f goldmane-6b4b7f4496-j4zb6 eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali7373238b47e [] [] }} ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.357 [INFO][5276] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.384 [INFO][5288] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" HandleID="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Workload="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.390 [INFO][5288] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" HandleID="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Workload="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ffa00), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"goldmane-6b4b7f4496-j4zb6", "timestamp":"2026-04-23 23:20:10.384784878 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40003a2f20)} Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.390 [INFO][5288] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.390 [INFO][5288] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.390 [INFO][5288] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.393 [INFO][5288] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.397 [INFO][5288] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.402 [INFO][5288] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.403 [INFO][5288] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.407 [INFO][5288] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.407 [INFO][5288] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.408 [INFO][5288] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.416 [INFO][5288] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.423 [INFO][5288] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.134/26] block=192.168.24.128/26 handle="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.423 [INFO][5288] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.134/26] handle="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.423 [INFO][5288] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:20:10.449985 containerd[1889]: 2026-04-23 23:20:10.423 [INFO][5288] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.134/26] IPv6=[] ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" HandleID="k8s-pod-network.666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Workload="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" Apr 23 23:20:10.450403 containerd[1889]: 2026-04-23 23:20:10.426 [INFO][5276] cni-plugin/k8s.go 418: Populated endpoint ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0", GenerateName:"goldmane-6b4b7f4496-", Namespace:"calico-system", SelfLink:"", UID:"b94f8cb1-0ece-4e1e-a59c-df0ce7bed609", ResourceVersion:"813", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"6b4b7f4496", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"goldmane-6b4b7f4496-j4zb6", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.24.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7373238b47e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:10.450403 containerd[1889]: 2026-04-23 23:20:10.426 [INFO][5276] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.134/32] ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" Apr 23 23:20:10.450403 containerd[1889]: 2026-04-23 23:20:10.426 [INFO][5276] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7373238b47e ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" Apr 23 23:20:10.450403 containerd[1889]: 2026-04-23 23:20:10.430 [INFO][5276] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" Apr 23 23:20:10.450403 containerd[1889]: 2026-04-23 23:20:10.433 [INFO][5276] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0", GenerateName:"goldmane-6b4b7f4496-", Namespace:"calico-system", SelfLink:"", UID:"b94f8cb1-0ece-4e1e-a59c-df0ce7bed609", ResourceVersion:"813", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"6b4b7f4496", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c", Pod:"goldmane-6b4b7f4496-j4zb6", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.24.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7373238b47e", MAC:"fe:57:af:00:f2:53", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:10.450403 containerd[1889]: 2026-04-23 23:20:10.447 [INFO][5276] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" Namespace="calico-system" Pod="goldmane-6b4b7f4496-j4zb6" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-goldmane--6b4b7f4496--j4zb6-eth0" Apr 23 23:20:10.499243 containerd[1889]: time="2026-04-23T23:20:10.499087065Z" level=info msg="connecting to shim 666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c" address="unix:///run/containerd/s/535bda46dc29c7da9acacd4179967b8be5a5b08051166be99c0eafd3164d0ee8" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:20:10.521756 systemd[1]: Started cri-containerd-666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c.scope - libcontainer container 666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c. Apr 23 23:20:10.540753 systemd-networkd[1474]: calie8f5711a37c: Gained IPv6LL Apr 23 23:20:10.563275 containerd[1889]: time="2026-04-23T23:20:10.563244944Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-6b4b7f4496-j4zb6,Uid:b94f8cb1-0ece-4e1e-a59c-df0ce7bed609,Namespace:calico-system,Attempt:0,} returns sandbox id \"666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c\"" Apr 23 23:20:10.988771 systemd-networkd[1474]: cali78b03fbf1af: Gained IPv6LL Apr 23 23:20:11.323345 containerd[1889]: time="2026-04-23T23:20:11.323087524Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-fkj7t,Uid:48f3e795-c1d9-4588-9444-c43e72073baf,Namespace:calico-system,Attempt:0,}" Apr 23 23:20:11.345267 containerd[1889]: time="2026-04-23T23:20:11.345227753Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:11.348278 containerd[1889]: time="2026-04-23T23:20:11.348254657Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.5: active requests=0, bytes read=46169343" Apr 23 23:20:11.351859 containerd[1889]: time="2026-04-23T23:20:11.351837647Z" level=info msg="ImageCreate event name:\"sha256:f3ba40f705afacb15a8a2f5b02c08a912321f045220eb8f8f1f5ca51f129741a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:11.358642 containerd[1889]: time="2026-04-23T23:20:11.358617580Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5fa7fb7e707d54479cd5d93cfe42352076b805f36560df457b53701d9e738d72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:11.359157 containerd[1889]: time="2026-04-23T23:20:11.358977634Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\" with image id \"sha256:f3ba40f705afacb15a8a2f5b02c08a912321f045220eb8f8f1f5ca51f129741a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5fa7fb7e707d54479cd5d93cfe42352076b805f36560df457b53701d9e738d72\", size \"48744950\" in 2.619189939s" Apr 23 23:20:11.359157 containerd[1889]: time="2026-04-23T23:20:11.359088543Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\" returns image reference \"sha256:f3ba40f705afacb15a8a2f5b02c08a912321f045220eb8f8f1f5ca51f129741a\"" Apr 23 23:20:11.361951 containerd[1889]: time="2026-04-23T23:20:11.361669901Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.5\"" Apr 23 23:20:11.384054 containerd[1889]: time="2026-04-23T23:20:11.384030859Z" level=info msg="CreateContainer within sandbox \"c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Apr 23 23:20:11.402219 containerd[1889]: time="2026-04-23T23:20:11.402190459Z" level=info msg="Container a9f8d8153406779c189499be7cdb536647b4b4b668a26c927ce5003ae7743016: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:11.419249 containerd[1889]: time="2026-04-23T23:20:11.419069600Z" level=info msg="CreateContainer within sandbox \"c4428368356302cc9a93bfa0b6e6d0071462106e37cb49b2d1f1122af7b16442\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"a9f8d8153406779c189499be7cdb536647b4b4b668a26c927ce5003ae7743016\"" Apr 23 23:20:11.420492 containerd[1889]: time="2026-04-23T23:20:11.420471896Z" level=info msg="StartContainer for \"a9f8d8153406779c189499be7cdb536647b4b4b668a26c927ce5003ae7743016\"" Apr 23 23:20:11.421535 containerd[1889]: time="2026-04-23T23:20:11.421515137Z" level=info msg="connecting to shim a9f8d8153406779c189499be7cdb536647b4b4b668a26c927ce5003ae7743016" address="unix:///run/containerd/s/56a17a9806fdbcf079b57de62b92931c0df50105ee9da5ad521e5d1eff134ae2" protocol=ttrpc version=3 Apr 23 23:20:11.440106 systemd-networkd[1474]: cali9e59394a359: Link UP Apr 23 23:20:11.440256 systemd-networkd[1474]: cali9e59394a359: Gained carrier Apr 23 23:20:11.440723 systemd[1]: Started cri-containerd-a9f8d8153406779c189499be7cdb536647b4b4b668a26c927ce5003ae7743016.scope - libcontainer container a9f8d8153406779c189499be7cdb536647b4b4b668a26c927ce5003ae7743016. Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.359 [INFO][5363] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0 calico-apiserver-57dbd4f578- calico-system 48f3e795-c1d9-4588-9444-c43e72073baf 814 0 2026-04-23 23:19:40 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:57dbd4f578 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f calico-apiserver-57dbd4f578-fkj7t eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali9e59394a359 [] [] }} ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.360 [INFO][5363] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.389 [INFO][5378] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" HandleID="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.394 [INFO][5378] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" HandleID="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000281910), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"calico-apiserver-57dbd4f578-fkj7t", "timestamp":"2026-04-23 23:20:11.389138902 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000346f20)} Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.394 [INFO][5378] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.394 [INFO][5378] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.394 [INFO][5378] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.397 [INFO][5378] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.408 [INFO][5378] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.412 [INFO][5378] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.414 [INFO][5378] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.416 [INFO][5378] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.416 [INFO][5378] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.418 [INFO][5378] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219 Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.423 [INFO][5378] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.433 [INFO][5378] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.135/26] block=192.168.24.128/26 handle="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.433 [INFO][5378] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.135/26] handle="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.433 [INFO][5378] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:20:11.463337 containerd[1889]: 2026-04-23 23:20:11.433 [INFO][5378] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.135/26] IPv6=[] ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" HandleID="k8s-pod-network.434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Workload="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" Apr 23 23:20:11.463715 containerd[1889]: 2026-04-23 23:20:11.435 [INFO][5363] cni-plugin/k8s.go 418: Populated endpoint ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0", GenerateName:"calico-apiserver-57dbd4f578-", Namespace:"calico-system", SelfLink:"", UID:"48f3e795-c1d9-4588-9444-c43e72073baf", ResourceVersion:"814", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57dbd4f578", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"calico-apiserver-57dbd4f578-fkj7t", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.24.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali9e59394a359", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:11.463715 containerd[1889]: 2026-04-23 23:20:11.435 [INFO][5363] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.135/32] ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" Apr 23 23:20:11.463715 containerd[1889]: 2026-04-23 23:20:11.435 [INFO][5363] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9e59394a359 ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" Apr 23 23:20:11.463715 containerd[1889]: 2026-04-23 23:20:11.447 [INFO][5363] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" Apr 23 23:20:11.463715 containerd[1889]: 2026-04-23 23:20:11.448 [INFO][5363] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0", GenerateName:"calico-apiserver-57dbd4f578-", Namespace:"calico-system", SelfLink:"", UID:"48f3e795-c1d9-4588-9444-c43e72073baf", ResourceVersion:"814", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57dbd4f578", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219", Pod:"calico-apiserver-57dbd4f578-fkj7t", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.24.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali9e59394a359", MAC:"62:fe:28:f8:52:62", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:11.463715 containerd[1889]: 2026-04-23 23:20:11.459 [INFO][5363] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" Namespace="calico-system" Pod="calico-apiserver-57dbd4f578-fkj7t" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-calico--apiserver--57dbd4f578--fkj7t-eth0" Apr 23 23:20:11.499417 containerd[1889]: time="2026-04-23T23:20:11.499389375Z" level=info msg="StartContainer for \"a9f8d8153406779c189499be7cdb536647b4b4b668a26c927ce5003ae7743016\" returns successfully" Apr 23 23:20:11.511992 containerd[1889]: time="2026-04-23T23:20:11.511858990Z" level=info msg="connecting to shim 434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219" address="unix:///run/containerd/s/4e8555b2ce850d69cbd944a06f639e9abc45f0e5fb1d317ae8c74a021f3cff5f" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:20:11.536647 systemd[1]: Started cri-containerd-434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219.scope - libcontainer container 434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219. Apr 23 23:20:11.595788 containerd[1889]: time="2026-04-23T23:20:11.595695409Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57dbd4f578-fkj7t,Uid:48f3e795-c1d9-4588-9444-c43e72073baf,Namespace:calico-system,Attempt:0,} returns sandbox id \"434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219\"" Apr 23 23:20:12.204714 systemd-networkd[1474]: cali7373238b47e: Gained IPv6LL Apr 23 23:20:12.322150 containerd[1889]: time="2026-04-23T23:20:12.322110592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zc5vb,Uid:9bd25f44-e1d9-4f63-84fb-09a335c9c974,Namespace:kube-system,Attempt:0,}" Apr 23 23:20:12.415500 systemd-networkd[1474]: cali255f15bd612: Link UP Apr 23 23:20:12.416697 systemd-networkd[1474]: cali255f15bd612: Gained carrier Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.355 [INFO][5484] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0 coredns-66bc5c9577- kube-system 9bd25f44-e1d9-4f63-84fb-09a335c9c974 811 0 2026-04-23 23:19:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459.2.4-n-8b75ae451f coredns-66bc5c9577-zc5vb eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali255f15bd612 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.355 [INFO][5484] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.377 [INFO][5496] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" HandleID="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Workload="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.383 [INFO][5496] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" HandleID="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Workload="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fbe90), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459.2.4-n-8b75ae451f", "pod":"coredns-66bc5c9577-zc5vb", "timestamp":"2026-04-23 23:20:12.37791902 +0000 UTC"}, Hostname:"ci-4459.2.4-n-8b75ae451f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400016a6e0)} Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.383 [INFO][5496] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.383 [INFO][5496] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.383 [INFO][5496] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.4-n-8b75ae451f' Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.385 [INFO][5496] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.389 [INFO][5496] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.392 [INFO][5496] ipam/ipam.go 526: Trying affinity for 192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.394 [INFO][5496] ipam/ipam.go 160: Attempting to load block cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.396 [INFO][5496] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.24.128/26 host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.396 [INFO][5496] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.24.128/26 handle="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.398 [INFO][5496] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25 Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.401 [INFO][5496] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.24.128/26 handle="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.410 [INFO][5496] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.24.136/26] block=192.168.24.128/26 handle="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.410 [INFO][5496] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.24.136/26] handle="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" host="ci-4459.2.4-n-8b75ae451f" Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.410 [INFO][5496] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 23 23:20:12.435260 containerd[1889]: 2026-04-23 23:20:12.410 [INFO][5496] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.24.136/26] IPv6=[] ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" HandleID="k8s-pod-network.42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Workload="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" Apr 23 23:20:12.436478 containerd[1889]: 2026-04-23 23:20:12.412 [INFO][5484] cni-plugin/k8s.go 418: Populated endpoint ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"9bd25f44-e1d9-4f63-84fb-09a335c9c974", ResourceVersion:"811", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"", Pod:"coredns-66bc5c9577-zc5vb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.24.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali255f15bd612", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:12.436478 containerd[1889]: 2026-04-23 23:20:12.412 [INFO][5484] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.24.136/32] ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" Apr 23 23:20:12.436478 containerd[1889]: 2026-04-23 23:20:12.412 [INFO][5484] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali255f15bd612 ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" Apr 23 23:20:12.436478 containerd[1889]: 2026-04-23 23:20:12.417 [INFO][5484] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" Apr 23 23:20:12.436478 containerd[1889]: 2026-04-23 23:20:12.417 [INFO][5484] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"9bd25f44-e1d9-4f63-84fb-09a335c9c974", ResourceVersion:"811", Generation:0, CreationTimestamp:time.Date(2026, time.April, 23, 23, 19, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.4-n-8b75ae451f", ContainerID:"42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25", Pod:"coredns-66bc5c9577-zc5vb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.24.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali255f15bd612", MAC:"6e:4a:88:dc:f3:30", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 23 23:20:12.436772 containerd[1889]: 2026-04-23 23:20:12.431 [INFO][5484] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" Namespace="kube-system" Pod="coredns-66bc5c9577-zc5vb" WorkloadEndpoint="ci--4459.2.4--n--8b75ae451f-k8s-coredns--66bc5c9577--zc5vb-eth0" Apr 23 23:20:12.505109 kubelet[3427]: I0423 23:20:12.504825 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-f94c58f4-b8pwc" podStartSLOduration=28.882646489 podStartE2EDuration="31.504795337s" podCreationTimestamp="2026-04-23 23:19:41 +0000 UTC" firstStartedPulling="2026-04-23 23:20:08.73913791 +0000 UTC m=+42.491259941" lastFinishedPulling="2026-04-23 23:20:11.361286758 +0000 UTC m=+45.113408789" observedRunningTime="2026-04-23 23:20:12.504310221 +0000 UTC m=+46.256432252" watchObservedRunningTime="2026-04-23 23:20:12.504795337 +0000 UTC m=+46.256917408" Apr 23 23:20:12.508392 containerd[1889]: time="2026-04-23T23:20:12.508062866Z" level=info msg="connecting to shim 42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25" address="unix:///run/containerd/s/9cd2854f99383fe27d88b1fea0b08163f85e29474dbbbfec863df25095ec76de" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:20:12.531153 systemd[1]: Started cri-containerd-42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25.scope - libcontainer container 42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25. Apr 23 23:20:12.569466 containerd[1889]: time="2026-04-23T23:20:12.569432250Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zc5vb,Uid:9bd25f44-e1d9-4f63-84fb-09a335c9c974,Namespace:kube-system,Attempt:0,} returns sandbox id \"42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25\"" Apr 23 23:20:12.581953 containerd[1889]: time="2026-04-23T23:20:12.581821797Z" level=info msg="CreateContainer within sandbox \"42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 23 23:20:12.607783 containerd[1889]: time="2026-04-23T23:20:12.607706215Z" level=info msg="Container c355fb1d643d566fbd4f24befae5be717e38e1d1901707330a5226bded503e39: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:12.623947 containerd[1889]: time="2026-04-23T23:20:12.623915194Z" level=info msg="CreateContainer within sandbox \"42591aa1f874a3c597e6dac3d4dc45f659aaf7efae040eb9d198cf866e085d25\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c355fb1d643d566fbd4f24befae5be717e38e1d1901707330a5226bded503e39\"" Apr 23 23:20:12.624731 containerd[1889]: time="2026-04-23T23:20:12.624701657Z" level=info msg="StartContainer for \"c355fb1d643d566fbd4f24befae5be717e38e1d1901707330a5226bded503e39\"" Apr 23 23:20:12.625486 containerd[1889]: time="2026-04-23T23:20:12.625456399Z" level=info msg="connecting to shim c355fb1d643d566fbd4f24befae5be717e38e1d1901707330a5226bded503e39" address="unix:///run/containerd/s/9cd2854f99383fe27d88b1fea0b08163f85e29474dbbbfec863df25095ec76de" protocol=ttrpc version=3 Apr 23 23:20:12.641705 systemd[1]: Started cri-containerd-c355fb1d643d566fbd4f24befae5be717e38e1d1901707330a5226bded503e39.scope - libcontainer container c355fb1d643d566fbd4f24befae5be717e38e1d1901707330a5226bded503e39. Apr 23 23:20:12.669299 containerd[1889]: time="2026-04-23T23:20:12.669230942Z" level=info msg="StartContainer for \"c355fb1d643d566fbd4f24befae5be717e38e1d1901707330a5226bded503e39\" returns successfully" Apr 23 23:20:12.716803 systemd-networkd[1474]: cali9e59394a359: Gained IPv6LL Apr 23 23:20:12.829880 containerd[1889]: time="2026-04-23T23:20:12.829451956Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:12.833031 containerd[1889]: time="2026-04-23T23:20:12.833000425Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.5: active requests=0, bytes read=7895994" Apr 23 23:20:12.836797 containerd[1889]: time="2026-04-23T23:20:12.836761462Z" level=info msg="ImageCreate event name:\"sha256:c84299759d8605dff0cc2ebb16a8c098e7266501883bb302cd068ecf668128a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:12.842099 containerd[1889]: time="2026-04-23T23:20:12.841970660Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e8a5b44388a309910946072582b1a1f283c52cf73e9825179235d934447c8b7d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:12.842396 containerd[1889]: time="2026-04-23T23:20:12.842365164Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.5\" with image id \"sha256:c84299759d8605dff0cc2ebb16a8c098e7266501883bb302cd068ecf668128a6\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e8a5b44388a309910946072582b1a1f283c52cf73e9825179235d934447c8b7d\", size \"10471633\" in 1.480325585s" Apr 23 23:20:12.842396 containerd[1889]: time="2026-04-23T23:20:12.842393197Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.5\" returns image reference \"sha256:c84299759d8605dff0cc2ebb16a8c098e7266501883bb302cd068ecf668128a6\"" Apr 23 23:20:12.844126 containerd[1889]: time="2026-04-23T23:20:12.844091328Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\"" Apr 23 23:20:12.853620 containerd[1889]: time="2026-04-23T23:20:12.851646812Z" level=info msg="CreateContainer within sandbox \"1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Apr 23 23:20:12.885730 containerd[1889]: time="2026-04-23T23:20:12.885706354Z" level=info msg="Container c377ac7b62c2373e5c16aec9d430e12a69490b65e962f17776bf7fbeb7864264: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:12.906148 containerd[1889]: time="2026-04-23T23:20:12.906110491Z" level=info msg="CreateContainer within sandbox \"1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"c377ac7b62c2373e5c16aec9d430e12a69490b65e962f17776bf7fbeb7864264\"" Apr 23 23:20:12.906795 containerd[1889]: time="2026-04-23T23:20:12.906766981Z" level=info msg="StartContainer for \"c377ac7b62c2373e5c16aec9d430e12a69490b65e962f17776bf7fbeb7864264\"" Apr 23 23:20:12.908114 containerd[1889]: time="2026-04-23T23:20:12.907954748Z" level=info msg="connecting to shim c377ac7b62c2373e5c16aec9d430e12a69490b65e962f17776bf7fbeb7864264" address="unix:///run/containerd/s/0c3e76de287c9643562696b5d38690f55425f44dc995173038811e32386e5baa" protocol=ttrpc version=3 Apr 23 23:20:12.925955 systemd[1]: Started cri-containerd-c377ac7b62c2373e5c16aec9d430e12a69490b65e962f17776bf7fbeb7864264.scope - libcontainer container c377ac7b62c2373e5c16aec9d430e12a69490b65e962f17776bf7fbeb7864264. Apr 23 23:20:13.001572 containerd[1889]: time="2026-04-23T23:20:13.001542153Z" level=info msg="StartContainer for \"c377ac7b62c2373e5c16aec9d430e12a69490b65e962f17776bf7fbeb7864264\" returns successfully" Apr 23 23:20:13.522570 kubelet[3427]: I0423 23:20:13.522512 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-zc5vb" podStartSLOduration=42.522498478 podStartE2EDuration="42.522498478s" podCreationTimestamp="2026-04-23 23:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:20:13.522112895 +0000 UTC m=+47.274234926" watchObservedRunningTime="2026-04-23 23:20:13.522498478 +0000 UTC m=+47.274620517" Apr 23 23:20:14.316723 systemd-networkd[1474]: cali255f15bd612: Gained IPv6LL Apr 23 23:20:15.746824 containerd[1889]: time="2026-04-23T23:20:15.746769032Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:15.749949 containerd[1889]: time="2026-04-23T23:20:15.749819779Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.5: active requests=0, bytes read=42617669" Apr 23 23:20:15.753368 containerd[1889]: time="2026-04-23T23:20:15.753343000Z" level=info msg="ImageCreate event name:\"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:15.758825 containerd[1889]: time="2026-04-23T23:20:15.758352045Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:78a11eeba8e8a02ecd6014bc8260180819ee7005f9eacb364b9595d1e4b166e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:15.758825 containerd[1889]: time="2026-04-23T23:20:15.758717482Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" with image id \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:78a11eeba8e8a02ecd6014bc8260180819ee7005f9eacb364b9595d1e4b166e1\", size \"45193324\" in 2.914600481s" Apr 23 23:20:15.758825 containerd[1889]: time="2026-04-23T23:20:15.758743483Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" returns image reference \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\"" Apr 23 23:20:15.760272 containerd[1889]: time="2026-04-23T23:20:15.760246988Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.5\"" Apr 23 23:20:15.777085 containerd[1889]: time="2026-04-23T23:20:15.777059438Z" level=info msg="CreateContainer within sandbox \"ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 23 23:20:15.814496 containerd[1889]: time="2026-04-23T23:20:15.814472169Z" level=info msg="Container 1d223424fb3135aeef0cb6f28805f8dbeb53f1b2d2c850356d365b50b4486b35: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:15.830687 containerd[1889]: time="2026-04-23T23:20:15.830662892Z" level=info msg="CreateContainer within sandbox \"ceee55781e7f0233b8732e0ba9ddf4bfbb117eb0d045bfc18ebc5d3e174875f3\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"1d223424fb3135aeef0cb6f28805f8dbeb53f1b2d2c850356d365b50b4486b35\"" Apr 23 23:20:15.831902 containerd[1889]: time="2026-04-23T23:20:15.831138238Z" level=info msg="StartContainer for \"1d223424fb3135aeef0cb6f28805f8dbeb53f1b2d2c850356d365b50b4486b35\"" Apr 23 23:20:15.832805 containerd[1889]: time="2026-04-23T23:20:15.832779532Z" level=info msg="connecting to shim 1d223424fb3135aeef0cb6f28805f8dbeb53f1b2d2c850356d365b50b4486b35" address="unix:///run/containerd/s/2fdf6c93fff9888e73586dfd72eafdd8516412a545e8f62716783c6dad686ae9" protocol=ttrpc version=3 Apr 23 23:20:15.851717 systemd[1]: Started cri-containerd-1d223424fb3135aeef0cb6f28805f8dbeb53f1b2d2c850356d365b50b4486b35.scope - libcontainer container 1d223424fb3135aeef0cb6f28805f8dbeb53f1b2d2c850356d365b50b4486b35. Apr 23 23:20:15.883752 containerd[1889]: time="2026-04-23T23:20:15.883727165Z" level=info msg="StartContainer for \"1d223424fb3135aeef0cb6f28805f8dbeb53f1b2d2c850356d365b50b4486b35\" returns successfully" Apr 23 23:20:16.539952 kubelet[3427]: I0423 23:20:16.539207 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-57dbd4f578-f7nhj" podStartSLOduration=30.353752577 podStartE2EDuration="36.539183239s" podCreationTimestamp="2026-04-23 23:19:40 +0000 UTC" firstStartedPulling="2026-04-23 23:20:09.574297155 +0000 UTC m=+43.326419194" lastFinishedPulling="2026-04-23 23:20:15.759727817 +0000 UTC m=+49.511849856" observedRunningTime="2026-04-23 23:20:16.538382009 +0000 UTC m=+50.290504088" watchObservedRunningTime="2026-04-23 23:20:16.539183239 +0000 UTC m=+50.291305278" Apr 23 23:20:17.525573 kubelet[3427]: I0423 23:20:17.525539 3427 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 23 23:20:18.161496 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3990707130.mount: Deactivated successfully. Apr 23 23:20:18.688503 containerd[1889]: time="2026-04-23T23:20:18.687989404Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:18.690975 containerd[1889]: time="2026-04-23T23:20:18.690951724Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.5: active requests=0, bytes read=48513326" Apr 23 23:20:18.693996 containerd[1889]: time="2026-04-23T23:20:18.693969798Z" level=info msg="ImageCreate event name:\"sha256:f556d75d96fa1483cf593e71a7d71a551e78433f43c12badd65e95187cd0fced\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:18.698305 containerd[1889]: time="2026-04-23T23:20:18.698281528Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:edfd1b6c377013f23afd5e76cb975b6cb59d1bc6554f79c0719d617f8dd0468e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:18.698937 containerd[1889]: time="2026-04-23T23:20:18.698626493Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.5\" with image id \"sha256:f556d75d96fa1483cf593e71a7d71a551e78433f43c12badd65e95187cd0fced\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:edfd1b6c377013f23afd5e76cb975b6cb59d1bc6554f79c0719d617f8dd0468e\", size \"48513172\" in 2.938353328s" Apr 23 23:20:18.698937 containerd[1889]: time="2026-04-23T23:20:18.698656183Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.5\" returns image reference \"sha256:f556d75d96fa1483cf593e71a7d71a551e78433f43c12badd65e95187cd0fced\"" Apr 23 23:20:18.700569 containerd[1889]: time="2026-04-23T23:20:18.700140383Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\"" Apr 23 23:20:18.706712 containerd[1889]: time="2026-04-23T23:20:18.706688438Z" level=info msg="CreateContainer within sandbox \"666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Apr 23 23:20:18.725787 containerd[1889]: time="2026-04-23T23:20:18.725708235Z" level=info msg="Container 4320216411070ac9179a3faa79c7ef9e054fcfeb24b63448a3cb58dcea740b58: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:18.745545 containerd[1889]: time="2026-04-23T23:20:18.745501661Z" level=info msg="CreateContainer within sandbox \"666a527e43f8242de1842879915b36ad3b9992967e790a3ee61b881d22daf64c\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"4320216411070ac9179a3faa79c7ef9e054fcfeb24b63448a3cb58dcea740b58\"" Apr 23 23:20:18.746420 containerd[1889]: time="2026-04-23T23:20:18.746217072Z" level=info msg="StartContainer for \"4320216411070ac9179a3faa79c7ef9e054fcfeb24b63448a3cb58dcea740b58\"" Apr 23 23:20:18.747400 containerd[1889]: time="2026-04-23T23:20:18.747378660Z" level=info msg="connecting to shim 4320216411070ac9179a3faa79c7ef9e054fcfeb24b63448a3cb58dcea740b58" address="unix:///run/containerd/s/535bda46dc29c7da9acacd4179967b8be5a5b08051166be99c0eafd3164d0ee8" protocol=ttrpc version=3 Apr 23 23:20:18.766708 systemd[1]: Started cri-containerd-4320216411070ac9179a3faa79c7ef9e054fcfeb24b63448a3cb58dcea740b58.scope - libcontainer container 4320216411070ac9179a3faa79c7ef9e054fcfeb24b63448a3cb58dcea740b58. Apr 23 23:20:18.799882 containerd[1889]: time="2026-04-23T23:20:18.799853655Z" level=info msg="StartContainer for \"4320216411070ac9179a3faa79c7ef9e054fcfeb24b63448a3cb58dcea740b58\" returns successfully" Apr 23 23:20:19.219645 containerd[1889]: time="2026-04-23T23:20:19.219164150Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:19.222872 containerd[1889]: time="2026-04-23T23:20:19.222848801Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.5: active requests=0, bytes read=77" Apr 23 23:20:19.224348 containerd[1889]: time="2026-04-23T23:20:19.224295872Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" with image id \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:78a11eeba8e8a02ecd6014bc8260180819ee7005f9eacb364b9595d1e4b166e1\", size \"45193324\" in 523.529833ms" Apr 23 23:20:19.224446 containerd[1889]: time="2026-04-23T23:20:19.224431613Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" returns image reference \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\"" Apr 23 23:20:19.225299 containerd[1889]: time="2026-04-23T23:20:19.225280101Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\"" Apr 23 23:20:19.232565 containerd[1889]: time="2026-04-23T23:20:19.232534631Z" level=info msg="CreateContainer within sandbox \"434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 23 23:20:19.257107 containerd[1889]: time="2026-04-23T23:20:19.256582881Z" level=info msg="Container cf2edc854bbc5c6fa10eb2b0926ba920747664ff085893aaebc296b5d27a3051: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:19.273992 containerd[1889]: time="2026-04-23T23:20:19.273969233Z" level=info msg="CreateContainer within sandbox \"434647a7bf5b7f6fa0f4418c7c7fd3ab4d1917c10595d30ca2f0a2aacf442219\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"cf2edc854bbc5c6fa10eb2b0926ba920747664ff085893aaebc296b5d27a3051\"" Apr 23 23:20:19.274801 containerd[1889]: time="2026-04-23T23:20:19.274698349Z" level=info msg="StartContainer for \"cf2edc854bbc5c6fa10eb2b0926ba920747664ff085893aaebc296b5d27a3051\"" Apr 23 23:20:19.276418 containerd[1889]: time="2026-04-23T23:20:19.276398669Z" level=info msg="connecting to shim cf2edc854bbc5c6fa10eb2b0926ba920747664ff085893aaebc296b5d27a3051" address="unix:///run/containerd/s/4e8555b2ce850d69cbd944a06f639e9abc45f0e5fb1d317ae8c74a021f3cff5f" protocol=ttrpc version=3 Apr 23 23:20:19.293747 systemd[1]: Started cri-containerd-cf2edc854bbc5c6fa10eb2b0926ba920747664ff085893aaebc296b5d27a3051.scope - libcontainer container cf2edc854bbc5c6fa10eb2b0926ba920747664ff085893aaebc296b5d27a3051. Apr 23 23:20:19.331832 containerd[1889]: time="2026-04-23T23:20:19.331804831Z" level=info msg="StartContainer for \"cf2edc854bbc5c6fa10eb2b0926ba920747664ff085893aaebc296b5d27a3051\" returns successfully" Apr 23 23:20:19.569833 kubelet[3427]: I0423 23:20:19.568806 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-57dbd4f578-fkj7t" podStartSLOduration=31.941977731 podStartE2EDuration="39.568792313s" podCreationTimestamp="2026-04-23 23:19:40 +0000 UTC" firstStartedPulling="2026-04-23 23:20:11.598373403 +0000 UTC m=+45.350495434" lastFinishedPulling="2026-04-23 23:20:19.225187985 +0000 UTC m=+52.977310016" observedRunningTime="2026-04-23 23:20:19.553297769 +0000 UTC m=+53.305419824" watchObservedRunningTime="2026-04-23 23:20:19.568792313 +0000 UTC m=+53.320914344" Apr 23 23:20:20.535949 kubelet[3427]: I0423 23:20:20.535915 3427 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 23 23:20:20.653347 containerd[1889]: time="2026-04-23T23:20:20.653130010Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:20.657889 containerd[1889]: time="2026-04-23T23:20:20.657861204Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5: active requests=0, bytes read=12456618" Apr 23 23:20:20.660670 containerd[1889]: time="2026-04-23T23:20:20.660645813Z" level=info msg="ImageCreate event name:\"sha256:a127885d176e495b4edc6e0c0309c6570e4d776444937bfdc565fac5a13d8b3f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:20.665153 containerd[1889]: time="2026-04-23T23:20:20.665120542Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:26849483b0c4d797a8ff818d988924bdf696996ca559c8c56b647aaaf70a448a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:20:20.665580 containerd[1889]: time="2026-04-23T23:20:20.665554727Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\" with image id \"sha256:a127885d176e495b4edc6e0c0309c6570e4d776444937bfdc565fac5a13d8b3f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:26849483b0c4d797a8ff818d988924bdf696996ca559c8c56b647aaaf70a448a\", size \"15032209\" in 1.440070362s" Apr 23 23:20:20.665615 containerd[1889]: time="2026-04-23T23:20:20.665583040Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\" returns image reference \"sha256:a127885d176e495b4edc6e0c0309c6570e4d776444937bfdc565fac5a13d8b3f\"" Apr 23 23:20:20.672583 containerd[1889]: time="2026-04-23T23:20:20.672554998Z" level=info msg="CreateContainer within sandbox \"1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Apr 23 23:20:20.707008 containerd[1889]: time="2026-04-23T23:20:20.706980473Z" level=info msg="Container 29649eaaeaa3b3d0a5c4dce97f00ef56eec147db1b2bdc6399261ae2f64ac1e7: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:20:20.709326 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3644311979.mount: Deactivated successfully. Apr 23 23:20:20.725882 containerd[1889]: time="2026-04-23T23:20:20.725789246Z" level=info msg="CreateContainer within sandbox \"1abfe0bee821dca39616abf86a9502d148c5b5af7a87949657ee9fd555c8c88d\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"29649eaaeaa3b3d0a5c4dce97f00ef56eec147db1b2bdc6399261ae2f64ac1e7\"" Apr 23 23:20:20.726240 containerd[1889]: time="2026-04-23T23:20:20.726222919Z" level=info msg="StartContainer for \"29649eaaeaa3b3d0a5c4dce97f00ef56eec147db1b2bdc6399261ae2f64ac1e7\"" Apr 23 23:20:20.727359 containerd[1889]: time="2026-04-23T23:20:20.727337417Z" level=info msg="connecting to shim 29649eaaeaa3b3d0a5c4dce97f00ef56eec147db1b2bdc6399261ae2f64ac1e7" address="unix:///run/containerd/s/0c3e76de287c9643562696b5d38690f55425f44dc995173038811e32386e5baa" protocol=ttrpc version=3 Apr 23 23:20:20.747717 systemd[1]: Started cri-containerd-29649eaaeaa3b3d0a5c4dce97f00ef56eec147db1b2bdc6399261ae2f64ac1e7.scope - libcontainer container 29649eaaeaa3b3d0a5c4dce97f00ef56eec147db1b2bdc6399261ae2f64ac1e7. Apr 23 23:20:20.803430 containerd[1889]: time="2026-04-23T23:20:20.803348764Z" level=info msg="StartContainer for \"29649eaaeaa3b3d0a5c4dce97f00ef56eec147db1b2bdc6399261ae2f64ac1e7\" returns successfully" Apr 23 23:20:21.393318 kubelet[3427]: I0423 23:20:21.393131 3427 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Apr 23 23:20:21.393318 kubelet[3427]: I0423 23:20:21.393167 3427 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Apr 23 23:20:21.558977 kubelet[3427]: I0423 23:20:21.558764 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-6b4b7f4496-j4zb6" podStartSLOduration=32.424236333 podStartE2EDuration="40.558750375s" podCreationTimestamp="2026-04-23 23:19:41 +0000 UTC" firstStartedPulling="2026-04-23 23:20:10.564781685 +0000 UTC m=+44.316903716" lastFinishedPulling="2026-04-23 23:20:18.699295719 +0000 UTC m=+52.451417758" observedRunningTime="2026-04-23 23:20:19.568669484 +0000 UTC m=+53.320791515" watchObservedRunningTime="2026-04-23 23:20:21.558750375 +0000 UTC m=+55.310872406" Apr 23 23:20:21.559850 kubelet[3427]: I0423 23:20:21.559813 3427 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-mlsd9" podStartSLOduration=28.709063334 podStartE2EDuration="40.559802326s" podCreationTimestamp="2026-04-23 23:19:41 +0000 UTC" firstStartedPulling="2026-04-23 23:20:08.815544242 +0000 UTC m=+42.567666273" lastFinishedPulling="2026-04-23 23:20:20.666283234 +0000 UTC m=+54.418405265" observedRunningTime="2026-04-23 23:20:21.558390929 +0000 UTC m=+55.310512960" watchObservedRunningTime="2026-04-23 23:20:21.559802326 +0000 UTC m=+55.311924389" Apr 23 23:20:32.469693 kubelet[3427]: I0423 23:20:32.469342 3427 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 23 23:20:41.548950 kubelet[3427]: I0423 23:20:41.548652 3427 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 23 23:20:44.391269 systemd[1]: Started sshd@7-10.0.0.13:22-50.85.169.122:36322.service - OpenSSH per-connection server daemon (50.85.169.122:36322). Apr 23 23:20:45.149938 sshd[5992]: Accepted publickey for core from 50.85.169.122 port 36322 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:20:45.151791 sshd-session[5992]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:20:45.155620 systemd-logind[1871]: New session 10 of user core. Apr 23 23:20:45.161727 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 23 23:20:45.629851 sshd[5995]: Connection closed by 50.85.169.122 port 36322 Apr 23 23:20:45.630304 sshd-session[5992]: pam_unix(sshd:session): session closed for user core Apr 23 23:20:45.633351 systemd[1]: sshd@7-10.0.0.13:22-50.85.169.122:36322.service: Deactivated successfully. Apr 23 23:20:45.635125 systemd[1]: session-10.scope: Deactivated successfully. Apr 23 23:20:45.637057 systemd-logind[1871]: Session 10 logged out. Waiting for processes to exit. Apr 23 23:20:45.638367 systemd-logind[1871]: Removed session 10. Apr 23 23:20:50.787224 systemd[1]: Started sshd@8-10.0.0.13:22-50.85.169.122:44018.service - OpenSSH per-connection server daemon (50.85.169.122:44018). Apr 23 23:20:51.561623 sshd[6008]: Accepted publickey for core from 50.85.169.122 port 44018 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:20:51.563817 sshd-session[6008]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:20:51.569436 systemd-logind[1871]: New session 11 of user core. Apr 23 23:20:51.575584 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 23 23:20:52.066000 sshd[6033]: Connection closed by 50.85.169.122 port 44018 Apr 23 23:20:52.066545 sshd-session[6008]: pam_unix(sshd:session): session closed for user core Apr 23 23:20:52.069284 systemd-logind[1871]: Session 11 logged out. Waiting for processes to exit. Apr 23 23:20:52.069383 systemd[1]: sshd@8-10.0.0.13:22-50.85.169.122:44018.service: Deactivated successfully. Apr 23 23:20:52.071347 systemd[1]: session-11.scope: Deactivated successfully. Apr 23 23:20:52.073459 systemd-logind[1871]: Removed session 11. Apr 23 23:20:57.220685 systemd[1]: Started sshd@9-10.0.0.13:22-50.85.169.122:44020.service - OpenSSH per-connection server daemon (50.85.169.122:44020). Apr 23 23:20:57.979319 sshd[6074]: Accepted publickey for core from 50.85.169.122 port 44020 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:20:57.980337 sshd-session[6074]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:20:57.984235 systemd-logind[1871]: New session 12 of user core. Apr 23 23:20:57.992716 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 23 23:20:58.471585 sshd[6077]: Connection closed by 50.85.169.122 port 44020 Apr 23 23:20:58.472153 sshd-session[6074]: pam_unix(sshd:session): session closed for user core Apr 23 23:20:58.475461 systemd[1]: sshd@9-10.0.0.13:22-50.85.169.122:44020.service: Deactivated successfully. Apr 23 23:20:58.477174 systemd[1]: session-12.scope: Deactivated successfully. Apr 23 23:20:58.478450 systemd-logind[1871]: Session 12 logged out. Waiting for processes to exit. Apr 23 23:20:58.479949 systemd-logind[1871]: Removed session 12. Apr 23 23:21:03.631380 systemd[1]: Started sshd@10-10.0.0.13:22-50.85.169.122:52420.service - OpenSSH per-connection server daemon (50.85.169.122:52420). Apr 23 23:21:04.406315 sshd[6115]: Accepted publickey for core from 50.85.169.122 port 52420 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:04.407373 sshd-session[6115]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:04.411665 systemd-logind[1871]: New session 13 of user core. Apr 23 23:21:04.416732 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 23 23:21:04.896694 sshd[6122]: Connection closed by 50.85.169.122 port 52420 Apr 23 23:21:04.897174 sshd-session[6115]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:04.900141 systemd[1]: sshd@10-10.0.0.13:22-50.85.169.122:52420.service: Deactivated successfully. Apr 23 23:21:04.903038 systemd[1]: session-13.scope: Deactivated successfully. Apr 23 23:21:04.904006 systemd-logind[1871]: Session 13 logged out. Waiting for processes to exit. Apr 23 23:21:04.905244 systemd-logind[1871]: Removed session 13. Apr 23 23:21:05.052949 systemd[1]: Started sshd@11-10.0.0.13:22-50.85.169.122:52426.service - OpenSSH per-connection server daemon (50.85.169.122:52426). Apr 23 23:21:05.822704 sshd[6135]: Accepted publickey for core from 50.85.169.122 port 52426 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:05.823708 sshd-session[6135]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:05.827331 systemd-logind[1871]: New session 14 of user core. Apr 23 23:21:05.833724 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 23 23:21:06.337819 sshd[6138]: Connection closed by 50.85.169.122 port 52426 Apr 23 23:21:06.338127 sshd-session[6135]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:06.343639 systemd-logind[1871]: Session 14 logged out. Waiting for processes to exit. Apr 23 23:21:06.344201 systemd[1]: sshd@11-10.0.0.13:22-50.85.169.122:52426.service: Deactivated successfully. Apr 23 23:21:06.345743 systemd[1]: session-14.scope: Deactivated successfully. Apr 23 23:21:06.347127 systemd-logind[1871]: Removed session 14. Apr 23 23:21:06.494323 systemd[1]: Started sshd@12-10.0.0.13:22-50.85.169.122:52436.service - OpenSSH per-connection server daemon (50.85.169.122:52436). Apr 23 23:21:07.264386 sshd[6148]: Accepted publickey for core from 50.85.169.122 port 52436 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:07.265717 sshd-session[6148]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:07.269282 systemd-logind[1871]: New session 15 of user core. Apr 23 23:21:07.274703 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 23 23:21:07.763304 sshd[6151]: Connection closed by 50.85.169.122 port 52436 Apr 23 23:21:07.766293 sshd-session[6148]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:07.770132 systemd[1]: sshd@12-10.0.0.13:22-50.85.169.122:52436.service: Deactivated successfully. Apr 23 23:21:07.773556 systemd[1]: session-15.scope: Deactivated successfully. Apr 23 23:21:07.775057 systemd-logind[1871]: Session 15 logged out. Waiting for processes to exit. Apr 23 23:21:07.778120 systemd-logind[1871]: Removed session 15. Apr 23 23:21:12.916572 systemd[1]: Started sshd@13-10.0.0.13:22-50.85.169.122:44054.service - OpenSSH per-connection server daemon (50.85.169.122:44054). Apr 23 23:21:13.662452 sshd[6185]: Accepted publickey for core from 50.85.169.122 port 44054 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:13.663893 sshd-session[6185]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:13.667400 systemd-logind[1871]: New session 16 of user core. Apr 23 23:21:13.672720 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 23 23:21:14.135768 sshd[6188]: Connection closed by 50.85.169.122 port 44054 Apr 23 23:21:14.136215 sshd-session[6185]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:14.139232 systemd[1]: sshd@13-10.0.0.13:22-50.85.169.122:44054.service: Deactivated successfully. Apr 23 23:21:14.140812 systemd[1]: session-16.scope: Deactivated successfully. Apr 23 23:21:14.141520 systemd-logind[1871]: Session 16 logged out. Waiting for processes to exit. Apr 23 23:21:14.144094 systemd-logind[1871]: Removed session 16. Apr 23 23:21:14.293825 systemd[1]: Started sshd@14-10.0.0.13:22-50.85.169.122:44064.service - OpenSSH per-connection server daemon (50.85.169.122:44064). Apr 23 23:21:15.068624 sshd[6200]: Accepted publickey for core from 50.85.169.122 port 44064 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:15.069546 sshd-session[6200]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:15.073337 systemd-logind[1871]: New session 17 of user core. Apr 23 23:21:15.077721 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 23 23:21:15.666280 sshd[6203]: Connection closed by 50.85.169.122 port 44064 Apr 23 23:21:15.666802 sshd-session[6200]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:15.670360 systemd[1]: sshd@14-10.0.0.13:22-50.85.169.122:44064.service: Deactivated successfully. Apr 23 23:21:15.672370 systemd[1]: session-17.scope: Deactivated successfully. Apr 23 23:21:15.673220 systemd-logind[1871]: Session 17 logged out. Waiting for processes to exit. Apr 23 23:21:15.674992 systemd-logind[1871]: Removed session 17. Apr 23 23:21:15.824612 systemd[1]: Started sshd@15-10.0.0.13:22-50.85.169.122:44080.service - OpenSSH per-connection server daemon (50.85.169.122:44080). Apr 23 23:21:16.594269 sshd[6213]: Accepted publickey for core from 50.85.169.122 port 44080 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:16.595334 sshd-session[6213]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:16.598823 systemd-logind[1871]: New session 18 of user core. Apr 23 23:21:16.604915 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 23 23:21:17.426702 sshd[6216]: Connection closed by 50.85.169.122 port 44080 Apr 23 23:21:17.426179 sshd-session[6213]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:17.429848 systemd-logind[1871]: Session 18 logged out. Waiting for processes to exit. Apr 23 23:21:17.430386 systemd[1]: sshd@15-10.0.0.13:22-50.85.169.122:44080.service: Deactivated successfully. Apr 23 23:21:17.432358 systemd[1]: session-18.scope: Deactivated successfully. Apr 23 23:21:17.434461 systemd-logind[1871]: Removed session 18. Apr 23 23:21:17.580079 systemd[1]: Started sshd@16-10.0.0.13:22-50.85.169.122:44086.service - OpenSSH per-connection server daemon (50.85.169.122:44086). Apr 23 23:21:18.334171 sshd[6236]: Accepted publickey for core from 50.85.169.122 port 44086 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:18.337228 sshd-session[6236]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:18.341255 systemd-logind[1871]: New session 19 of user core. Apr 23 23:21:18.349809 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 23 23:21:18.894924 sshd[6241]: Connection closed by 50.85.169.122 port 44086 Apr 23 23:21:18.896185 sshd-session[6236]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:18.900527 systemd[1]: sshd@16-10.0.0.13:22-50.85.169.122:44086.service: Deactivated successfully. Apr 23 23:21:18.902246 systemd[1]: session-19.scope: Deactivated successfully. Apr 23 23:21:18.903120 systemd-logind[1871]: Session 19 logged out. Waiting for processes to exit. Apr 23 23:21:18.904959 systemd-logind[1871]: Removed session 19. Apr 23 23:21:19.047148 systemd[1]: Started sshd@17-10.0.0.13:22-50.85.169.122:44094.service - OpenSSH per-connection server daemon (50.85.169.122:44094). Apr 23 23:21:19.797767 sshd[6252]: Accepted publickey for core from 50.85.169.122 port 44094 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:19.799007 sshd-session[6252]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:19.802858 systemd-logind[1871]: New session 20 of user core. Apr 23 23:21:19.808723 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 23 23:21:20.265571 sshd[6255]: Connection closed by 50.85.169.122 port 44094 Apr 23 23:21:20.266788 sshd-session[6252]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:20.269792 systemd-logind[1871]: Session 20 logged out. Waiting for processes to exit. Apr 23 23:21:20.270306 systemd[1]: sshd@17-10.0.0.13:22-50.85.169.122:44094.service: Deactivated successfully. Apr 23 23:21:20.273179 systemd[1]: session-20.scope: Deactivated successfully. Apr 23 23:21:20.274966 systemd-logind[1871]: Removed session 20. Apr 23 23:21:25.418946 systemd[1]: Started sshd@18-10.0.0.13:22-50.85.169.122:56332.service - OpenSSH per-connection server daemon (50.85.169.122:56332). Apr 23 23:21:26.163908 sshd[6297]: Accepted publickey for core from 50.85.169.122 port 56332 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:26.164960 sshd-session[6297]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:26.169650 systemd-logind[1871]: New session 21 of user core. Apr 23 23:21:26.174719 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 23 23:21:26.636096 sshd[6322]: Connection closed by 50.85.169.122 port 56332 Apr 23 23:21:26.636648 sshd-session[6297]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:26.640550 systemd-logind[1871]: Session 21 logged out. Waiting for processes to exit. Apr 23 23:21:26.640845 systemd[1]: sshd@18-10.0.0.13:22-50.85.169.122:56332.service: Deactivated successfully. Apr 23 23:21:26.642518 systemd[1]: session-21.scope: Deactivated successfully. Apr 23 23:21:26.644045 systemd-logind[1871]: Removed session 21. Apr 23 23:21:31.792250 systemd[1]: Started sshd@19-10.0.0.13:22-50.85.169.122:41376.service - OpenSSH per-connection server daemon (50.85.169.122:41376). Apr 23 23:21:32.541492 sshd[6337]: Accepted publickey for core from 50.85.169.122 port 41376 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:32.543012 sshd-session[6337]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:32.548656 systemd-logind[1871]: New session 22 of user core. Apr 23 23:21:32.554782 systemd[1]: Started session-22.scope - Session 22 of User core. Apr 23 23:21:33.015417 sshd[6367]: Connection closed by 50.85.169.122 port 41376 Apr 23 23:21:33.014559 sshd-session[6337]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:33.017672 systemd[1]: sshd@19-10.0.0.13:22-50.85.169.122:41376.service: Deactivated successfully. Apr 23 23:21:33.019321 systemd[1]: session-22.scope: Deactivated successfully. Apr 23 23:21:33.020039 systemd-logind[1871]: Session 22 logged out. Waiting for processes to exit. Apr 23 23:21:33.021116 systemd-logind[1871]: Removed session 22. Apr 23 23:21:38.167424 systemd[1]: Started sshd@20-10.0.0.13:22-50.85.169.122:41386.service - OpenSSH per-connection server daemon (50.85.169.122:41386). Apr 23 23:21:38.911437 sshd[6401]: Accepted publickey for core from 50.85.169.122 port 41386 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:38.912509 sshd-session[6401]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:38.916428 systemd-logind[1871]: New session 23 of user core. Apr 23 23:21:38.920706 systemd[1]: Started session-23.scope - Session 23 of User core. Apr 23 23:21:39.385690 sshd[6404]: Connection closed by 50.85.169.122 port 41386 Apr 23 23:21:39.386431 sshd-session[6401]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:39.389418 systemd-logind[1871]: Session 23 logged out. Waiting for processes to exit. Apr 23 23:21:39.389741 systemd[1]: sshd@20-10.0.0.13:22-50.85.169.122:41386.service: Deactivated successfully. Apr 23 23:21:39.391544 systemd[1]: session-23.scope: Deactivated successfully. Apr 23 23:21:39.394328 systemd-logind[1871]: Removed session 23. Apr 23 23:21:44.533779 systemd[1]: Started sshd@21-10.0.0.13:22-50.85.169.122:48164.service - OpenSSH per-connection server daemon (50.85.169.122:48164). Apr 23 23:21:45.266020 sshd[6438]: Accepted publickey for core from 50.85.169.122 port 48164 ssh2: RSA SHA256:OE/BzpIjp/Jg1G36L5zUqHa7NG/Z9l5Fwb+VInZbsf0 Apr 23 23:21:45.267413 sshd-session[6438]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:21:45.271238 systemd-logind[1871]: New session 24 of user core. Apr 23 23:21:45.274703 systemd[1]: Started session-24.scope - Session 24 of User core. Apr 23 23:21:45.733628 sshd[6441]: Connection closed by 50.85.169.122 port 48164 Apr 23 23:21:45.734784 sshd-session[6438]: pam_unix(sshd:session): session closed for user core Apr 23 23:21:45.737864 systemd-logind[1871]: Session 24 logged out. Waiting for processes to exit. Apr 23 23:21:45.738483 systemd[1]: sshd@21-10.0.0.13:22-50.85.169.122:48164.service: Deactivated successfully. Apr 23 23:21:45.741026 systemd[1]: session-24.scope: Deactivated successfully. Apr 23 23:21:45.742589 systemd-logind[1871]: Removed session 24.