Mar 2 12:48:06.041026 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd490] Mar 2 12:48:06.041043 kernel: Linux version 6.12.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Mon Mar 2 10:44:26 -00 2026 Mar 2 12:48:06.041049 kernel: KASLR enabled Mar 2 12:48:06.041053 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Mar 2 12:48:06.041057 kernel: printk: legacy bootconsole [pl11] enabled Mar 2 12:48:06.041061 kernel: efi: EFI v2.7 by EDK II Mar 2 12:48:06.041067 kernel: efi: ACPI 2.0=0x3f979018 SMBIOS=0x3f8a0000 SMBIOS 3.0=0x3f880000 MEMATTR=0x3e89c018 RNG=0x3f979998 MEMRESERVE=0x3db83598 Mar 2 12:48:06.041071 kernel: random: crng init done Mar 2 12:48:06.041075 kernel: secureboot: Secure boot disabled Mar 2 12:48:06.041079 kernel: ACPI: Early table checksum verification disabled Mar 2 12:48:06.041082 kernel: ACPI: RSDP 0x000000003F979018 000024 (v02 VRTUAL) Mar 2 12:48:06.041086 kernel: ACPI: XSDT 0x000000003F979F18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041090 kernel: ACPI: FACP 0x000000003F979C18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041094 kernel: ACPI: DSDT 0x000000003F95A018 01E046 (v02 MSFTVM DSDT01 00000001 INTL 20230628) Mar 2 12:48:06.041100 kernel: ACPI: DBG2 0x000000003F979B18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041104 kernel: ACPI: GTDT 0x000000003F979D98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041109 kernel: ACPI: OEM0 0x000000003F979098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041113 kernel: ACPI: SPCR 0x000000003F979A98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041117 kernel: ACPI: APIC 0x000000003F979818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041122 kernel: ACPI: SRAT 0x000000003F979198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041126 kernel: ACPI: PPTT 0x000000003F979418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Mar 2 12:48:06.041131 kernel: ACPI: BGRT 0x000000003F979E98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:48:06.041135 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Mar 2 12:48:06.041139 kernel: ACPI: Use ACPI SPCR as default console: Yes Mar 2 12:48:06.041143 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Mar 2 12:48:06.041147 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] hotplug Mar 2 12:48:06.041152 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] hotplug Mar 2 12:48:06.041156 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Mar 2 12:48:06.041160 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Mar 2 12:48:06.041164 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Mar 2 12:48:06.041169 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Mar 2 12:48:06.041174 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Mar 2 12:48:06.041178 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Mar 2 12:48:06.041182 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Mar 2 12:48:06.041186 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Mar 2 12:48:06.041190 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Mar 2 12:48:06.041195 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x1bfffffff] -> [mem 0x00000000-0x1bfffffff] Mar 2 12:48:06.041199 kernel: NODE_DATA(0) allocated [mem 0x1bf7ffa00-0x1bf806fff] Mar 2 12:48:06.041203 kernel: Zone ranges: Mar 2 12:48:06.041207 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Mar 2 12:48:06.041214 kernel: DMA32 empty Mar 2 12:48:06.041218 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Mar 2 12:48:06.041223 kernel: Device empty Mar 2 12:48:06.041227 kernel: Movable zone start for each node Mar 2 12:48:06.041231 kernel: Early memory node ranges Mar 2 12:48:06.041236 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Mar 2 12:48:06.041241 kernel: node 0: [mem 0x0000000000824000-0x000000003f38ffff] Mar 2 12:48:06.041245 kernel: node 0: [mem 0x000000003f390000-0x000000003f93ffff] Mar 2 12:48:06.041250 kernel: node 0: [mem 0x000000003f940000-0x000000003f9effff] Mar 2 12:48:06.041254 kernel: node 0: [mem 0x000000003f9f0000-0x000000003fdeffff] Mar 2 12:48:06.041258 kernel: node 0: [mem 0x000000003fdf0000-0x000000003fffffff] Mar 2 12:48:06.041263 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Mar 2 12:48:06.041267 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Mar 2 12:48:06.041271 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Mar 2 12:48:06.041276 kernel: cma: Reserved 16 MiB at 0x000000003ca00000 on node -1 Mar 2 12:48:06.041280 kernel: psci: probing for conduit method from ACPI. Mar 2 12:48:06.041284 kernel: psci: PSCIv1.3 detected in firmware. Mar 2 12:48:06.041289 kernel: psci: Using standard PSCI v0.2 function IDs Mar 2 12:48:06.041294 kernel: psci: MIGRATE_INFO_TYPE not supported. Mar 2 12:48:06.041298 kernel: psci: SMC Calling Convention v1.4 Mar 2 12:48:06.041303 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Mar 2 12:48:06.041307 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Mar 2 12:48:06.041311 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Mar 2 12:48:06.041316 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Mar 2 12:48:06.041320 kernel: pcpu-alloc: [0] 0 [0] 1 Mar 2 12:48:06.041325 kernel: Detected PIPT I-cache on CPU0 Mar 2 12:48:06.041329 kernel: CPU features: detected: Address authentication (architected QARMA5 algorithm) Mar 2 12:48:06.041334 kernel: CPU features: detected: GIC system register CPU interface Mar 2 12:48:06.041338 kernel: CPU features: detected: Spectre-v4 Mar 2 12:48:06.041342 kernel: CPU features: detected: Spectre-BHB Mar 2 12:48:06.041347 kernel: CPU features: kernel page table isolation forced ON by KASLR Mar 2 12:48:06.041352 kernel: CPU features: detected: Kernel page table isolation (KPTI) Mar 2 12:48:06.041356 kernel: CPU features: detected: ARM erratum 2067961 or 2054223 Mar 2 12:48:06.041361 kernel: CPU features: detected: SSBS not fully self-synchronizing Mar 2 12:48:06.041365 kernel: alternatives: applying boot alternatives Mar 2 12:48:06.041386 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=75d5e65dbf56ddb5ea243beb025fcfbdb9b2a65e9b1b7d7db3d24aed3f0a168f Mar 2 12:48:06.041391 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 2 12:48:06.041396 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 2 12:48:06.041400 kernel: Fallback order for Node 0: 0 Mar 2 12:48:06.041404 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1048540 Mar 2 12:48:06.041410 kernel: Policy zone: Normal Mar 2 12:48:06.041414 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 2 12:48:06.041418 kernel: software IO TLB: area num 2. Mar 2 12:48:06.041423 kernel: software IO TLB: mapped [mem 0x0000000035900000-0x0000000039900000] (64MB) Mar 2 12:48:06.041428 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 2 12:48:06.041432 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 2 12:48:06.041437 kernel: rcu: RCU event tracing is enabled. Mar 2 12:48:06.041441 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 2 12:48:06.041446 kernel: Trampoline variant of Tasks RCU enabled. Mar 2 12:48:06.041450 kernel: Tracing variant of Tasks RCU enabled. Mar 2 12:48:06.041455 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 2 12:48:06.041459 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 2 12:48:06.041465 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 2 12:48:06.041469 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 2 12:48:06.041474 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 2 12:48:06.041478 kernel: GICv3: 960 SPIs implemented Mar 2 12:48:06.041482 kernel: GICv3: 0 Extended SPIs implemented Mar 2 12:48:06.041487 kernel: Root IRQ handler: gic_handle_irq Mar 2 12:48:06.041491 kernel: GICv3: GICv3 features: 16 PPIs, RSS Mar 2 12:48:06.041495 kernel: GICv3: GICD_CTRL.DS=0, SCR_EL3.FIQ=0 Mar 2 12:48:06.041500 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Mar 2 12:48:06.041504 kernel: ITS: No ITS available, not enabling LPIs Mar 2 12:48:06.041508 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 2 12:48:06.041514 kernel: arch_timer: cp15 timer(s) running at 1000.00MHz (virt). Mar 2 12:48:06.041518 kernel: clocksource: arch_sys_counter: mask: 0x1fffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 2 12:48:06.041523 kernel: sched_clock: 61 bits at 1000MHz, resolution 1ns, wraps every 4398046511103ns Mar 2 12:48:06.041527 kernel: Console: colour dummy device 80x25 Mar 2 12:48:06.041532 kernel: printk: legacy console [tty1] enabled Mar 2 12:48:06.041536 kernel: ACPI: Core revision 20240827 Mar 2 12:48:06.041541 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 2000.00 BogoMIPS (lpj=1000000) Mar 2 12:48:06.041546 kernel: pid_max: default: 32768 minimum: 301 Mar 2 12:48:06.041550 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Mar 2 12:48:06.041555 kernel: landlock: Up and running. Mar 2 12:48:06.041560 kernel: SELinux: Initializing. Mar 2 12:48:06.041565 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 12:48:06.041569 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 12:48:06.041574 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0xa0000e, misc 0x31e1 Mar 2 12:48:06.041579 kernel: Hyper-V: Host Build 10.0.26102.1212-1-0 Mar 2 12:48:06.041586 kernel: Hyper-V: enabling crash_kexec_post_notifiers Mar 2 12:48:06.041592 kernel: rcu: Hierarchical SRCU implementation. Mar 2 12:48:06.041597 kernel: rcu: Max phase no-delay instances is 400. Mar 2 12:48:06.041601 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Mar 2 12:48:06.041606 kernel: Remapping and enabling EFI services. Mar 2 12:48:06.041611 kernel: smp: Bringing up secondary CPUs ... Mar 2 12:48:06.041616 kernel: Detected PIPT I-cache on CPU1 Mar 2 12:48:06.041621 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Mar 2 12:48:06.041626 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd490] Mar 2 12:48:06.041631 kernel: smp: Brought up 1 node, 2 CPUs Mar 2 12:48:06.041635 kernel: SMP: Total of 2 processors activated. Mar 2 12:48:06.041640 kernel: CPU: All CPU(s) started at EL1 Mar 2 12:48:06.041646 kernel: CPU features: detected: 32-bit EL0 Support Mar 2 12:48:06.041650 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Mar 2 12:48:06.041655 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Mar 2 12:48:06.041660 kernel: CPU features: detected: Common not Private translations Mar 2 12:48:06.041665 kernel: CPU features: detected: CRC32 instructions Mar 2 12:48:06.041670 kernel: CPU features: detected: Generic authentication (architected QARMA5 algorithm) Mar 2 12:48:06.041675 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Mar 2 12:48:06.041679 kernel: CPU features: detected: LSE atomic instructions Mar 2 12:48:06.041684 kernel: CPU features: detected: Privileged Access Never Mar 2 12:48:06.041689 kernel: CPU features: detected: Speculation barrier (SB) Mar 2 12:48:06.041694 kernel: CPU features: detected: TLB range maintenance instructions Mar 2 12:48:06.041699 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Mar 2 12:48:06.041704 kernel: CPU features: detected: Scalable Vector Extension Mar 2 12:48:06.041709 kernel: alternatives: applying system-wide alternatives Mar 2 12:48:06.041713 kernel: CPU features: detected: Hardware dirty bit management on CPU0-1 Mar 2 12:48:06.041718 kernel: SVE: maximum available vector length 16 bytes per vector Mar 2 12:48:06.041723 kernel: SVE: default vector length 16 bytes per vector Mar 2 12:48:06.041728 kernel: Memory: 3952828K/4194160K available (11200K kernel code, 2458K rwdata, 9088K rodata, 39552K init, 1038K bss, 220144K reserved, 16384K cma-reserved) Mar 2 12:48:06.041734 kernel: devtmpfs: initialized Mar 2 12:48:06.041738 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 2 12:48:06.041743 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 2 12:48:06.041748 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Mar 2 12:48:06.041753 kernel: 0 pages in range for non-PLT usage Mar 2 12:48:06.041757 kernel: 508400 pages in range for PLT usage Mar 2 12:48:06.041762 kernel: pinctrl core: initialized pinctrl subsystem Mar 2 12:48:06.041767 kernel: SMBIOS 3.1.0 present. Mar 2 12:48:06.041772 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 06/10/2025 Mar 2 12:48:06.041777 kernel: DMI: Memory slots populated: 2/2 Mar 2 12:48:06.041782 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 2 12:48:06.041787 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 2 12:48:06.041792 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 2 12:48:06.041796 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 2 12:48:06.041801 kernel: audit: initializing netlink subsys (disabled) Mar 2 12:48:06.041806 kernel: audit: type=2000 audit(0.059:1): state=initialized audit_enabled=0 res=1 Mar 2 12:48:06.041811 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 2 12:48:06.041816 kernel: cpuidle: using governor menu Mar 2 12:48:06.041821 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 2 12:48:06.041826 kernel: ASID allocator initialised with 32768 entries Mar 2 12:48:06.041831 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 2 12:48:06.041835 kernel: Serial: AMBA PL011 UART driver Mar 2 12:48:06.041840 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 2 12:48:06.041845 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Mar 2 12:48:06.041850 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Mar 2 12:48:06.041854 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Mar 2 12:48:06.041862 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 2 12:48:06.041867 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Mar 2 12:48:06.041871 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Mar 2 12:48:06.041876 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Mar 2 12:48:06.041881 kernel: ACPI: Added _OSI(Module Device) Mar 2 12:48:06.041885 kernel: ACPI: Added _OSI(Processor Device) Mar 2 12:48:06.041890 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 2 12:48:06.041895 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 2 12:48:06.041900 kernel: ACPI: Interpreter enabled Mar 2 12:48:06.041905 kernel: ACPI: Using GIC for interrupt routing Mar 2 12:48:06.041910 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Mar 2 12:48:06.041915 kernel: printk: legacy console [ttyAMA0] enabled Mar 2 12:48:06.041919 kernel: printk: legacy bootconsole [pl11] disabled Mar 2 12:48:06.041924 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Mar 2 12:48:06.041929 kernel: ACPI: CPU0 has been hot-added Mar 2 12:48:06.041934 kernel: ACPI: CPU1 has been hot-added Mar 2 12:48:06.041938 kernel: iommu: Default domain type: Translated Mar 2 12:48:06.041943 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 2 12:48:06.041948 kernel: efivars: Registered efivars operations Mar 2 12:48:06.041953 kernel: vgaarb: loaded Mar 2 12:48:06.041958 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 2 12:48:06.041963 kernel: VFS: Disk quotas dquot_6.6.0 Mar 2 12:48:06.041967 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 2 12:48:06.041972 kernel: pnp: PnP ACPI init Mar 2 12:48:06.041977 kernel: pnp: PnP ACPI: found 0 devices Mar 2 12:48:06.041981 kernel: NET: Registered PF_INET protocol family Mar 2 12:48:06.041986 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 2 12:48:06.041991 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 2 12:48:06.041997 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 2 12:48:06.042001 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 2 12:48:06.042006 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 2 12:48:06.042011 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 2 12:48:06.042016 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 12:48:06.042021 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 12:48:06.042025 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 2 12:48:06.042030 kernel: PCI: CLS 0 bytes, default 64 Mar 2 12:48:06.042035 kernel: kvm [1]: HYP mode not available Mar 2 12:48:06.042040 kernel: Initialise system trusted keyrings Mar 2 12:48:06.042045 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 2 12:48:06.042050 kernel: Key type asymmetric registered Mar 2 12:48:06.042055 kernel: Asymmetric key parser 'x509' registered Mar 2 12:48:06.042059 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Mar 2 12:48:06.042064 kernel: io scheduler mq-deadline registered Mar 2 12:48:06.042069 kernel: io scheduler kyber registered Mar 2 12:48:06.042074 kernel: io scheduler bfq registered Mar 2 12:48:06.042078 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 2 12:48:06.042084 kernel: thunder_xcv, ver 1.0 Mar 2 12:48:06.042088 kernel: thunder_bgx, ver 1.0 Mar 2 12:48:06.042093 kernel: nicpf, ver 1.0 Mar 2 12:48:06.042098 kernel: nicvf, ver 1.0 Mar 2 12:48:06.042195 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 2 12:48:06.042245 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-03-02T12:48:05 UTC (1772455685) Mar 2 12:48:06.042252 kernel: efifb: probing for efifb Mar 2 12:48:06.042258 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Mar 2 12:48:06.042263 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Mar 2 12:48:06.042268 kernel: efifb: scrolling: redraw Mar 2 12:48:06.042272 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 2 12:48:06.042277 kernel: Console: switching to colour frame buffer device 128x48 Mar 2 12:48:06.042282 kernel: fb0: EFI VGA frame buffer device Mar 2 12:48:06.042287 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Mar 2 12:48:06.042291 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 2 12:48:06.042296 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Mar 2 12:48:06.042302 kernel: NET: Registered PF_INET6 protocol family Mar 2 12:48:06.042307 kernel: watchdog: NMI not fully supported Mar 2 12:48:06.042311 kernel: watchdog: Hard watchdog permanently disabled Mar 2 12:48:06.042316 kernel: Segment Routing with IPv6 Mar 2 12:48:06.042321 kernel: In-situ OAM (IOAM) with IPv6 Mar 2 12:48:06.042326 kernel: NET: Registered PF_PACKET protocol family Mar 2 12:48:06.042330 kernel: Key type dns_resolver registered Mar 2 12:48:06.042335 kernel: registered taskstats version 1 Mar 2 12:48:06.042340 kernel: Loading compiled-in X.509 certificates Mar 2 12:48:06.042345 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.74-flatcar: 03854795d80c6b1eedd5f94f64a67d19428ce88e' Mar 2 12:48:06.042350 kernel: Demotion targets for Node 0: null Mar 2 12:48:06.042355 kernel: Key type .fscrypt registered Mar 2 12:48:06.042360 kernel: Key type fscrypt-provisioning registered Mar 2 12:48:06.042364 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 2 12:48:06.044401 kernel: ima: Allocated hash algorithm: sha1 Mar 2 12:48:06.044409 kernel: ima: No architecture policies found Mar 2 12:48:06.044415 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 2 12:48:06.044420 kernel: clk: Disabling unused clocks Mar 2 12:48:06.044425 kernel: PM: genpd: Disabling unused power domains Mar 2 12:48:06.044433 kernel: Warning: unable to open an initial console. Mar 2 12:48:06.044438 kernel: Freeing unused kernel memory: 39552K Mar 2 12:48:06.044443 kernel: Run /init as init process Mar 2 12:48:06.044447 kernel: with arguments: Mar 2 12:48:06.044452 kernel: /init Mar 2 12:48:06.044457 kernel: with environment: Mar 2 12:48:06.044462 kernel: HOME=/ Mar 2 12:48:06.044467 kernel: TERM=linux Mar 2 12:48:06.044473 systemd[1]: Successfully made /usr/ read-only. Mar 2 12:48:06.044481 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 12:48:06.044487 systemd[1]: Detected virtualization microsoft. Mar 2 12:48:06.044492 systemd[1]: Detected architecture arm64. Mar 2 12:48:06.044497 systemd[1]: Running in initrd. Mar 2 12:48:06.044503 systemd[1]: No hostname configured, using default hostname. Mar 2 12:48:06.044508 systemd[1]: Hostname set to . Mar 2 12:48:06.044513 systemd[1]: Initializing machine ID from random generator. Mar 2 12:48:06.044520 systemd[1]: Queued start job for default target initrd.target. Mar 2 12:48:06.044525 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 12:48:06.044530 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 12:48:06.044536 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 2 12:48:06.044541 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 12:48:06.044547 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 2 12:48:06.044552 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 2 12:48:06.044559 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 2 12:48:06.044565 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 2 12:48:06.044570 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 12:48:06.044575 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 12:48:06.044580 systemd[1]: Reached target paths.target - Path Units. Mar 2 12:48:06.044586 systemd[1]: Reached target slices.target - Slice Units. Mar 2 12:48:06.044591 systemd[1]: Reached target swap.target - Swaps. Mar 2 12:48:06.044596 systemd[1]: Reached target timers.target - Timer Units. Mar 2 12:48:06.044602 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 12:48:06.044607 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 12:48:06.044613 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 2 12:48:06.044618 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 2 12:48:06.044623 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 12:48:06.044629 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 12:48:06.044634 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 12:48:06.044639 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 12:48:06.044645 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 2 12:48:06.044650 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 12:48:06.044656 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 2 12:48:06.044661 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Mar 2 12:48:06.044666 systemd[1]: Starting systemd-fsck-usr.service... Mar 2 12:48:06.044672 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 12:48:06.044677 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 12:48:06.044700 systemd-journald[225]: Collecting audit messages is disabled. Mar 2 12:48:06.044714 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:48:06.044720 systemd-journald[225]: Journal started Mar 2 12:48:06.044735 systemd-journald[225]: Runtime Journal (/run/log/journal/57d08fa9ad2e47c9b9ae9494065199c5) is 8M, max 78.3M, 70.3M free. Mar 2 12:48:06.053175 systemd-modules-load[227]: Inserted module 'overlay' Mar 2 12:48:06.068627 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 12:48:06.068652 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 2 12:48:06.077387 kernel: Bridge firewalling registered Mar 2 12:48:06.078226 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 2 12:48:06.079139 systemd-modules-load[227]: Inserted module 'br_netfilter' Mar 2 12:48:06.091525 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 12:48:06.096698 systemd[1]: Finished systemd-fsck-usr.service. Mar 2 12:48:06.105680 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 12:48:06.114377 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:06.123402 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 2 12:48:06.142740 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 12:48:06.147313 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 12:48:06.163452 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 12:48:06.179593 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:48:06.188481 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 12:48:06.190322 systemd-tmpfiles[247]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Mar 2 12:48:06.194599 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 12:48:06.205397 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 12:48:06.223669 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 2 12:48:06.239888 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 12:48:06.247181 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 12:48:06.265350 dracut-cmdline[262]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=75d5e65dbf56ddb5ea243beb025fcfbdb9b2a65e9b1b7d7db3d24aed3f0a168f Mar 2 12:48:06.290750 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 12:48:06.317164 systemd-resolved[263]: Positive Trust Anchors: Mar 2 12:48:06.320285 systemd-resolved[263]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 12:48:06.333569 kernel: SCSI subsystem initialized Mar 2 12:48:06.333587 kernel: Loading iSCSI transport class v2.0-870. Mar 2 12:48:06.320306 systemd-resolved[263]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 12:48:06.322002 systemd-resolved[263]: Defaulting to hostname 'linux'. Mar 2 12:48:06.322624 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 12:48:06.363249 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 12:48:06.383477 kernel: iscsi: registered transport (tcp) Mar 2 12:48:06.394570 kernel: iscsi: registered transport (qla4xxx) Mar 2 12:48:06.394578 kernel: QLogic iSCSI HBA Driver Mar 2 12:48:06.406597 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 12:48:06.429516 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 12:48:06.435932 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 12:48:06.485401 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 2 12:48:06.491477 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 2 12:48:06.554383 kernel: raid6: neonx8 gen() 18540 MB/s Mar 2 12:48:06.571375 kernel: raid6: neonx4 gen() 18546 MB/s Mar 2 12:48:06.590376 kernel: raid6: neonx2 gen() 17066 MB/s Mar 2 12:48:06.610377 kernel: raid6: neonx1 gen() 15019 MB/s Mar 2 12:48:06.629376 kernel: raid6: int64x8 gen() 10527 MB/s Mar 2 12:48:06.648459 kernel: raid6: int64x4 gen() 10618 MB/s Mar 2 12:48:06.669385 kernel: raid6: int64x2 gen() 8973 MB/s Mar 2 12:48:06.689496 kernel: raid6: int64x1 gen() 7048 MB/s Mar 2 12:48:06.689507 kernel: raid6: using algorithm neonx4 gen() 18546 MB/s Mar 2 12:48:06.711384 kernel: raid6: .... xor() 15146 MB/s, rmw enabled Mar 2 12:48:06.711390 kernel: raid6: using neon recovery algorithm Mar 2 12:48:06.720029 kernel: xor: measuring software checksum speed Mar 2 12:48:06.720039 kernel: 8regs : 28591 MB/sec Mar 2 12:48:06.722484 kernel: 32regs : 28812 MB/sec Mar 2 12:48:06.724876 kernel: arm64_neon : 37587 MB/sec Mar 2 12:48:06.727751 kernel: xor: using function: arm64_neon (37587 MB/sec) Mar 2 12:48:06.765383 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 2 12:48:06.770272 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 2 12:48:06.779509 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 12:48:06.814059 systemd-udevd[474]: Using default interface naming scheme 'v255'. Mar 2 12:48:06.816935 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 12:48:06.830053 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 2 12:48:06.858858 dracut-pre-trigger[485]: rd.md=0: removing MD RAID activation Mar 2 12:48:06.878245 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 12:48:06.883877 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 12:48:06.923443 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 12:48:06.935827 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 2 12:48:06.994381 kernel: hv_vmbus: Vmbus version:5.3 Mar 2 12:48:07.009822 kernel: hv_vmbus: registering driver hyperv_keyboard Mar 2 12:48:07.009854 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Mar 2 12:48:07.015386 kernel: hv_vmbus: registering driver hv_storvsc Mar 2 12:48:07.016548 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:48:07.050935 kernel: scsi host0: storvsc_host_t Mar 2 12:48:07.060986 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Mar 2 12:48:07.061012 kernel: scsi host1: storvsc_host_t Mar 2 12:48:07.066475 kernel: hv_vmbus: registering driver hid_hyperv Mar 2 12:48:07.066491 kernel: pps_core: LinuxPPS API ver. 1 registered Mar 2 12:48:07.066498 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 5 Mar 2 12:48:07.066612 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Mar 2 12:48:07.066620 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Mar 2 12:48:07.066626 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Mar 2 12:48:07.066694 kernel: hv_vmbus: registering driver hv_netvsc Mar 2 12:48:07.016720 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:07.062144 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:48:07.094420 kernel: PTP clock support registered Mar 2 12:48:07.094436 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Mar 2 12:48:07.109474 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Mar 2 12:48:07.112889 kernel: sd 0:0:0:0: [sda] Write Protect is off Mar 2 12:48:07.113150 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Mar 2 12:48:07.116045 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Mar 2 12:48:07.116351 kernel: hv_utils: Registering HyperV Utility Driver Mar 2 12:48:07.116364 kernel: hv_vmbus: registering driver hv_utils Mar 2 12:48:07.076646 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:48:06.683736 kernel: hv_utils: Heartbeat IC version 3.0 Mar 2 12:48:06.688913 kernel: hv_utils: Shutdown IC version 3.2 Mar 2 12:48:06.688924 kernel: hv_utils: TimeSync IC version 4.0 Mar 2 12:48:06.688930 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 2 12:48:06.688935 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Mar 2 12:48:06.689029 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Mar 2 12:48:06.689096 systemd-journald[225]: Time jumped backwards, rotating. Mar 2 12:48:06.689122 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 2 12:48:07.098935 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:48:06.695439 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Mar 2 12:48:07.099001 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:07.113648 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 12:48:07.114926 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:48:06.671420 systemd-resolved[263]: Clock change detected. Flushing caches. Mar 2 12:48:06.727740 kernel: hv_netvsc 7ced8db6-fce0-7ced-8db6-fce07ced8db6 eth0: VF slot 1 added Mar 2 12:48:06.727857 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#181 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 2 12:48:06.713569 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:06.751317 kernel: hv_vmbus: registering driver hv_pci Mar 2 12:48:06.751341 kernel: hv_pci f2e3e33f-6b1f-418d-9447-7e06da2c965e: PCI VMBus probing: Using version 0x10004 Mar 2 12:48:06.762289 kernel: hv_pci f2e3e33f-6b1f-418d-9447-7e06da2c965e: PCI host bridge to bus 6b1f:00 Mar 2 12:48:06.762406 kernel: pci_bus 6b1f:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Mar 2 12:48:06.762487 kernel: pci_bus 6b1f:00: No busn resource found for root bus, will use [bus 00-ff] Mar 2 12:48:06.778202 kernel: pci 6b1f:00:02.0: [15b3:101a] type 00 class 0x020000 PCIe Endpoint Mar 2 12:48:06.778244 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#243 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 2 12:48:06.789539 kernel: pci 6b1f:00:02.0: BAR 0 [mem 0xfc0000000-0xfc00fffff 64bit pref] Mar 2 12:48:06.793533 kernel: pci 6b1f:00:02.0: enabling Extended Tags Mar 2 12:48:06.808506 kernel: pci 6b1f:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 6b1f:00:02.0 (capable of 252.048 Gb/s with 16.0 GT/s PCIe x16 link) Mar 2 12:48:06.817069 kernel: pci_bus 6b1f:00: busn_res: [bus 00-ff] end is updated to 00 Mar 2 12:48:06.817188 kernel: pci 6b1f:00:02.0: BAR 0 [mem 0xfc0000000-0xfc00fffff 64bit pref]: assigned Mar 2 12:48:06.874984 kernel: mlx5_core 6b1f:00:02.0: enabling device (0000 -> 0002) Mar 2 12:48:06.882736 kernel: mlx5_core 6b1f:00:02.0: PTM is not supported by PCIe Mar 2 12:48:06.882876 kernel: mlx5_core 6b1f:00:02.0: firmware version: 16.30.5026 Mar 2 12:48:07.056286 kernel: hv_netvsc 7ced8db6-fce0-7ced-8db6-fce07ced8db6 eth0: VF registering: eth1 Mar 2 12:48:07.056459 kernel: mlx5_core 6b1f:00:02.0 eth1: joined to eth0 Mar 2 12:48:07.062043 kernel: mlx5_core 6b1f:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Mar 2 12:48:07.071512 kernel: mlx5_core 6b1f:00:02.0 enP27423s1: renamed from eth1 Mar 2 12:48:07.150896 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Mar 2 12:48:07.222650 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Mar 2 12:48:07.244952 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Mar 2 12:48:07.286918 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Mar 2 12:48:07.292127 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Mar 2 12:48:07.301952 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 2 12:48:07.313449 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 12:48:07.322243 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 12:48:07.331598 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 12:48:07.340359 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 2 12:48:07.364036 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 2 12:48:07.383781 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 2 12:48:07.397730 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 2 12:48:08.413910 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 2 12:48:08.413959 disk-uuid[667]: The operation has completed successfully. Mar 2 12:48:08.485024 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 2 12:48:08.487621 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 2 12:48:08.511315 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 2 12:48:08.526430 sh[826]: Success Mar 2 12:48:08.558231 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 2 12:48:08.558265 kernel: device-mapper: uevent: version 1.0.3 Mar 2 12:48:08.563016 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Mar 2 12:48:08.571517 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Mar 2 12:48:08.812924 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 2 12:48:08.819662 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 2 12:48:08.834519 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 2 12:48:08.856509 kernel: BTRFS: device fsid da6bd89d-75a6-483e-9a3e-89df5ed9b6c4 devid 1 transid 36 /dev/mapper/usr (254:0) scanned by mount (844) Mar 2 12:48:08.866764 kernel: BTRFS info (device dm-0): first mount of filesystem da6bd89d-75a6-483e-9a3e-89df5ed9b6c4 Mar 2 12:48:08.866774 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:48:09.210607 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Mar 2 12:48:09.210678 kernel: BTRFS info (device dm-0 state E): enabling free space tree Mar 2 12:48:09.328369 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 2 12:48:09.332450 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Mar 2 12:48:09.339456 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 2 12:48:09.340077 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 2 12:48:09.361358 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 2 12:48:09.387535 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (867) Mar 2 12:48:09.397227 kernel: BTRFS info (device sda6): first mount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:48:09.397254 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:48:09.419973 kernel: BTRFS info (device sda6): turning on async discard Mar 2 12:48:09.420001 kernel: BTRFS info (device sda6): enabling free space tree Mar 2 12:48:09.427536 kernel: BTRFS info (device sda6): last unmount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:48:09.428715 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 2 12:48:09.435605 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 2 12:48:09.473053 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 12:48:09.478932 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 12:48:09.511130 systemd-networkd[1013]: lo: Link UP Mar 2 12:48:09.511139 systemd-networkd[1013]: lo: Gained carrier Mar 2 12:48:09.511848 systemd-networkd[1013]: Enumeration completed Mar 2 12:48:09.513604 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 12:48:09.516787 systemd-networkd[1013]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:48:09.516790 systemd-networkd[1013]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 12:48:09.520773 systemd[1]: Reached target network.target - Network. Mar 2 12:48:09.590514 kernel: mlx5_core 6b1f:00:02.0 enP27423s1: Link up Mar 2 12:48:09.622667 kernel: hv_netvsc 7ced8db6-fce0-7ced-8db6-fce07ced8db6 eth0: Data path switched to VF: enP27423s1 Mar 2 12:48:09.622869 systemd-networkd[1013]: enP27423s1: Link UP Mar 2 12:48:09.625760 systemd-networkd[1013]: eth0: Link UP Mar 2 12:48:09.625831 systemd-networkd[1013]: eth0: Gained carrier Mar 2 12:48:09.625839 systemd-networkd[1013]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:48:09.640608 systemd-networkd[1013]: enP27423s1: Gained carrier Mar 2 12:48:09.652516 systemd-networkd[1013]: eth0: DHCPv4 address 10.200.20.38/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 2 12:48:10.686425 ignition[960]: Ignition 2.22.0 Mar 2 12:48:10.686442 ignition[960]: Stage: fetch-offline Mar 2 12:48:10.692429 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 12:48:10.686546 ignition[960]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:48:10.686553 ignition[960]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:48:10.702956 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 2 12:48:10.686629 ignition[960]: parsed url from cmdline: "" Mar 2 12:48:10.686632 ignition[960]: no config URL provided Mar 2 12:48:10.686635 ignition[960]: reading system config file "/usr/lib/ignition/user.ign" Mar 2 12:48:10.686640 ignition[960]: no config at "/usr/lib/ignition/user.ign" Mar 2 12:48:10.686643 ignition[960]: failed to fetch config: resource requires networking Mar 2 12:48:10.686756 ignition[960]: Ignition finished successfully Mar 2 12:48:10.735427 ignition[1023]: Ignition 2.22.0 Mar 2 12:48:10.735432 ignition[1023]: Stage: fetch Mar 2 12:48:10.735676 ignition[1023]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:48:10.735683 ignition[1023]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:48:10.735745 ignition[1023]: parsed url from cmdline: "" Mar 2 12:48:10.735748 ignition[1023]: no config URL provided Mar 2 12:48:10.735751 ignition[1023]: reading system config file "/usr/lib/ignition/user.ign" Mar 2 12:48:10.735758 ignition[1023]: no config at "/usr/lib/ignition/user.ign" Mar 2 12:48:10.735773 ignition[1023]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Mar 2 12:48:10.827977 ignition[1023]: GET result: OK Mar 2 12:48:10.828026 ignition[1023]: config has been read from IMDS userdata Mar 2 12:48:10.828047 ignition[1023]: parsing config with SHA512: 9d373857248a975cad790f5a846f9290a54de3ea06e8ac52d25d013ab42cb2748745c1288744a6895d8207c7336ca5497c1103f414065938ea17ae468306411b Mar 2 12:48:10.831128 unknown[1023]: fetched base config from "system" Mar 2 12:48:10.831439 ignition[1023]: fetch: fetch complete Mar 2 12:48:10.831140 unknown[1023]: fetched base config from "system" Mar 2 12:48:10.831442 ignition[1023]: fetch: fetch passed Mar 2 12:48:10.831143 unknown[1023]: fetched user config from "azure" Mar 2 12:48:10.831482 ignition[1023]: Ignition finished successfully Mar 2 12:48:10.835573 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 2 12:48:10.843365 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 2 12:48:10.878589 ignition[1030]: Ignition 2.22.0 Mar 2 12:48:10.878598 ignition[1030]: Stage: kargs Mar 2 12:48:10.878802 ignition[1030]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:48:10.882106 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 2 12:48:10.878810 ignition[1030]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:48:10.886692 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 2 12:48:10.879262 ignition[1030]: kargs: kargs passed Mar 2 12:48:10.879292 ignition[1030]: Ignition finished successfully Mar 2 12:48:10.915976 ignition[1036]: Ignition 2.22.0 Mar 2 12:48:10.915993 ignition[1036]: Stage: disks Mar 2 12:48:10.918674 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 2 12:48:10.916158 ignition[1036]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:48:10.924706 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 2 12:48:10.916165 ignition[1036]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:48:10.932382 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 2 12:48:10.916736 ignition[1036]: disks: disks passed Mar 2 12:48:10.940908 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 12:48:10.916766 ignition[1036]: Ignition finished successfully Mar 2 12:48:10.949247 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 12:48:10.957497 systemd[1]: Reached target basic.target - Basic System. Mar 2 12:48:10.966317 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 2 12:48:11.048252 systemd-fsck[1044]: ROOT: clean, 15/7326000 files, 477845/7359488 blocks Mar 2 12:48:11.056032 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 2 12:48:11.062290 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 2 12:48:11.294514 kernel: EXT4-fs (sda9): mounted filesystem 6408ffd3-d563-490c-803b-1f4582ee0319 r/w with ordered data mode. Quota mode: none. Mar 2 12:48:11.294933 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 2 12:48:11.298604 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 2 12:48:11.327361 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 12:48:11.341027 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 2 12:48:11.356259 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 2 12:48:11.369725 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (1058) Mar 2 12:48:11.364889 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 2 12:48:11.392370 kernel: BTRFS info (device sda6): first mount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:48:11.392385 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:48:11.364914 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 12:48:11.380335 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 2 12:48:11.398605 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 2 12:48:11.420433 kernel: BTRFS info (device sda6): turning on async discard Mar 2 12:48:11.420458 kernel: BTRFS info (device sda6): enabling free space tree Mar 2 12:48:11.422193 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 12:48:11.634626 systemd-networkd[1013]: eth0: Gained IPv6LL Mar 2 12:48:12.041092 coreos-metadata[1060]: Mar 02 12:48:12.041 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 2 12:48:12.049445 coreos-metadata[1060]: Mar 02 12:48:12.049 INFO Fetch successful Mar 2 12:48:12.053398 coreos-metadata[1060]: Mar 02 12:48:12.053 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Mar 2 12:48:12.062096 coreos-metadata[1060]: Mar 02 12:48:12.062 INFO Fetch successful Mar 2 12:48:12.066314 coreos-metadata[1060]: Mar 02 12:48:12.063 INFO wrote hostname ci-4459.2.101-47783670b7 to /sysroot/etc/hostname Mar 2 12:48:12.066707 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 2 12:48:12.274302 initrd-setup-root[1088]: cut: /sysroot/etc/passwd: No such file or directory Mar 2 12:48:12.311046 initrd-setup-root[1095]: cut: /sysroot/etc/group: No such file or directory Mar 2 12:48:12.318357 initrd-setup-root[1102]: cut: /sysroot/etc/shadow: No such file or directory Mar 2 12:48:12.340311 initrd-setup-root[1109]: cut: /sysroot/etc/gshadow: No such file or directory Mar 2 12:48:13.123561 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 2 12:48:13.128785 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 2 12:48:13.145065 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 2 12:48:13.156011 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 2 12:48:13.166517 kernel: BTRFS info (device sda6): last unmount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:48:13.183165 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 2 12:48:13.192827 ignition[1177]: INFO : Ignition 2.22.0 Mar 2 12:48:13.192827 ignition[1177]: INFO : Stage: mount Mar 2 12:48:13.199882 ignition[1177]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 12:48:13.199882 ignition[1177]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:48:13.199882 ignition[1177]: INFO : mount: mount passed Mar 2 12:48:13.199882 ignition[1177]: INFO : Ignition finished successfully Mar 2 12:48:13.200172 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 2 12:48:13.208744 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 2 12:48:13.231590 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 12:48:13.259504 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (1188) Mar 2 12:48:13.275844 kernel: BTRFS info (device sda6): first mount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:48:13.275872 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:48:13.285449 kernel: BTRFS info (device sda6): turning on async discard Mar 2 12:48:13.285463 kernel: BTRFS info (device sda6): enabling free space tree Mar 2 12:48:13.286945 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 12:48:13.317883 ignition[1205]: INFO : Ignition 2.22.0 Mar 2 12:48:13.317883 ignition[1205]: INFO : Stage: files Mar 2 12:48:13.323610 ignition[1205]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 12:48:13.323610 ignition[1205]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:48:13.323610 ignition[1205]: DEBUG : files: compiled without relabeling support, skipping Mar 2 12:48:13.347371 ignition[1205]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 2 12:48:13.347371 ignition[1205]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 2 12:48:13.406589 ignition[1205]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 2 12:48:13.411983 ignition[1205]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 2 12:48:13.411983 ignition[1205]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 2 12:48:13.407885 unknown[1205]: wrote ssh authorized keys file for user: core Mar 2 12:48:13.453822 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 2 12:48:13.461444 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Mar 2 12:48:13.494441 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 2 12:48:13.629804 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 2 12:48:13.629804 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 2 12:48:13.644004 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 2 12:48:13.644004 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 2 12:48:13.644004 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 2 12:48:13.644004 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 12:48:13.644004 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 12:48:13.644004 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 12:48:13.644004 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 12:48:13.692546 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 12:48:13.692546 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 12:48:13.692546 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-arm64.raw" Mar 2 12:48:13.692546 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-arm64.raw" Mar 2 12:48:13.692546 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-arm64.raw" Mar 2 12:48:13.692546 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.35.1-arm64.raw: attempt #1 Mar 2 12:48:14.166512 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Mar 2 12:48:15.008089 ignition[1205]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-arm64.raw" Mar 2 12:48:15.008089 ignition[1205]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Mar 2 12:48:15.055087 ignition[1205]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 12:48:15.067712 ignition[1205]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 12:48:15.067712 ignition[1205]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Mar 2 12:48:15.082627 ignition[1205]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Mar 2 12:48:15.082627 ignition[1205]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Mar 2 12:48:15.082627 ignition[1205]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 2 12:48:15.082627 ignition[1205]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 2 12:48:15.082627 ignition[1205]: INFO : files: files passed Mar 2 12:48:15.082627 ignition[1205]: INFO : Ignition finished successfully Mar 2 12:48:15.076679 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 2 12:48:15.087454 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 2 12:48:15.109862 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 2 12:48:15.127630 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 2 12:48:15.134703 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 2 12:48:15.170125 initrd-setup-root-after-ignition[1235]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 12:48:15.170125 initrd-setup-root-after-ignition[1235]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 2 12:48:15.183210 initrd-setup-root-after-ignition[1239]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 12:48:15.183837 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 12:48:15.194363 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 2 12:48:15.204502 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 2 12:48:15.238726 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 2 12:48:15.238811 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 2 12:48:15.248503 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 2 12:48:15.257439 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 2 12:48:15.265565 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 2 12:48:15.266124 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 2 12:48:15.298392 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 12:48:15.304861 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 2 12:48:15.330335 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 2 12:48:15.335560 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 12:48:15.344895 systemd[1]: Stopped target timers.target - Timer Units. Mar 2 12:48:15.353634 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 2 12:48:15.353709 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 12:48:15.366040 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 2 12:48:15.374788 systemd[1]: Stopped target basic.target - Basic System. Mar 2 12:48:15.382315 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 2 12:48:15.390321 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 12:48:15.399430 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 2 12:48:15.408580 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Mar 2 12:48:15.417790 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 2 12:48:15.426628 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 12:48:15.435876 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 2 12:48:15.444734 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 2 12:48:15.452902 systemd[1]: Stopped target swap.target - Swaps. Mar 2 12:48:15.460140 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 2 12:48:15.460228 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 2 12:48:15.471589 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 2 12:48:15.480142 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 12:48:15.489255 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 2 12:48:15.493811 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 12:48:15.499120 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 2 12:48:15.499196 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 2 12:48:15.512772 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 2 12:48:15.512861 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 12:48:15.518342 systemd[1]: ignition-files.service: Deactivated successfully. Mar 2 12:48:15.518408 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 2 12:48:15.526519 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 2 12:48:15.526582 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 2 12:48:15.536392 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 2 12:48:15.564654 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 2 12:48:15.582927 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 2 12:48:15.583034 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 12:48:15.605956 ignition[1259]: INFO : Ignition 2.22.0 Mar 2 12:48:15.605956 ignition[1259]: INFO : Stage: umount Mar 2 12:48:15.605956 ignition[1259]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 12:48:15.605956 ignition[1259]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:48:15.605956 ignition[1259]: INFO : umount: umount passed Mar 2 12:48:15.605956 ignition[1259]: INFO : Ignition finished successfully Mar 2 12:48:15.589843 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 2 12:48:15.591026 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 12:48:15.610623 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 2 12:48:15.610701 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 2 12:48:15.619147 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 2 12:48:15.619228 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 2 12:48:15.626215 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 2 12:48:15.626253 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 2 12:48:15.636418 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 2 12:48:15.636462 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 2 12:48:15.649266 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 2 12:48:15.649309 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 2 12:48:15.657281 systemd[1]: Stopped target network.target - Network. Mar 2 12:48:15.664787 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 2 12:48:15.664838 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 12:48:15.673878 systemd[1]: Stopped target paths.target - Path Units. Mar 2 12:48:15.683722 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 2 12:48:15.687510 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 12:48:15.695929 systemd[1]: Stopped target slices.target - Slice Units. Mar 2 12:48:15.704285 systemd[1]: Stopped target sockets.target - Socket Units. Mar 2 12:48:15.715467 systemd[1]: iscsid.socket: Deactivated successfully. Mar 2 12:48:15.715512 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 12:48:15.722917 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 2 12:48:15.722946 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 12:48:15.730862 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 2 12:48:15.730906 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 2 12:48:15.738419 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 2 12:48:15.738445 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 2 12:48:15.747199 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 2 12:48:15.754356 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 2 12:48:15.764068 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 2 12:48:15.768370 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 2 12:48:15.768469 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 2 12:48:15.783625 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Mar 2 12:48:15.783823 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 2 12:48:15.783906 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 2 12:48:15.797276 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Mar 2 12:48:15.797694 systemd[1]: Stopped target network-pre.target - Preparation for Network. Mar 2 12:48:15.949702 kernel: hv_netvsc 7ced8db6-fce0-7ced-8db6-fce07ced8db6 eth0: Data path switched from VF: enP27423s1 Mar 2 12:48:15.805058 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 2 12:48:15.805104 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 2 12:48:15.813772 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 2 12:48:15.826131 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 2 12:48:15.826185 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 12:48:15.835064 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 2 12:48:15.835100 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:48:15.846893 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 2 12:48:15.846929 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 2 12:48:15.851718 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 2 12:48:15.851747 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 12:48:15.864640 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 12:48:15.872835 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 2 12:48:15.872886 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Mar 2 12:48:15.883567 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 2 12:48:15.886451 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 2 12:48:15.894512 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 2 12:48:15.894550 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 2 12:48:15.903407 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 2 12:48:15.903540 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 12:48:15.913540 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 2 12:48:15.913600 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 2 12:48:15.921604 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 2 12:48:15.921628 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 12:48:15.929752 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 2 12:48:15.929785 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 2 12:48:15.949832 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 2 12:48:15.949900 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 2 12:48:15.962371 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 2 12:48:15.962900 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 12:48:15.976715 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 2 12:48:15.992080 systemd[1]: systemd-network-generator.service: Deactivated successfully. Mar 2 12:48:15.992130 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 12:48:16.004773 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 2 12:48:16.004810 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 12:48:16.010484 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Mar 2 12:48:16.010537 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 12:48:16.019851 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 2 12:48:16.019887 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 12:48:16.192939 systemd-journald[225]: Received SIGTERM from PID 1 (systemd). Mar 2 12:48:16.029243 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:48:16.029280 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:16.044926 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Mar 2 12:48:16.044964 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Mar 2 12:48:16.044985 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 2 12:48:16.045010 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 12:48:16.045224 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 2 12:48:16.045300 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 2 12:48:16.053561 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 2 12:48:16.053625 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 2 12:48:16.064856 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 2 12:48:16.073087 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 2 12:48:16.094789 systemd[1]: Switching root. Mar 2 12:48:16.250246 systemd-journald[225]: Journal stopped Mar 2 12:48:20.671038 kernel: SELinux: policy capability network_peer_controls=1 Mar 2 12:48:20.671055 kernel: SELinux: policy capability open_perms=1 Mar 2 12:48:20.671063 kernel: SELinux: policy capability extended_socket_class=1 Mar 2 12:48:20.671068 kernel: SELinux: policy capability always_check_network=0 Mar 2 12:48:20.671073 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 2 12:48:20.671079 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 2 12:48:20.671085 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 2 12:48:20.671091 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 2 12:48:20.671096 kernel: SELinux: policy capability userspace_initial_context=0 Mar 2 12:48:20.671101 kernel: audit: type=1403 audit(1772455697.278:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 2 12:48:20.671107 systemd[1]: Successfully loaded SELinux policy in 191.625ms. Mar 2 12:48:20.671116 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 4.346ms. Mar 2 12:48:20.671122 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 12:48:20.671128 systemd[1]: Detected virtualization microsoft. Mar 2 12:48:20.671135 systemd[1]: Detected architecture arm64. Mar 2 12:48:20.671141 systemd[1]: Detected first boot. Mar 2 12:48:20.671148 systemd[1]: Hostname set to . Mar 2 12:48:20.671154 systemd[1]: Initializing machine ID from random generator. Mar 2 12:48:20.671160 zram_generator::config[1301]: No configuration found. Mar 2 12:48:20.671166 kernel: NET: Registered PF_VSOCK protocol family Mar 2 12:48:20.671171 systemd[1]: Populated /etc with preset unit settings. Mar 2 12:48:20.671178 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Mar 2 12:48:20.671183 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 2 12:48:20.671190 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 2 12:48:20.671196 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 2 12:48:20.671202 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 2 12:48:20.671208 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 2 12:48:20.671214 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 2 12:48:20.671220 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 2 12:48:20.671226 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 2 12:48:20.671233 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 2 12:48:20.671239 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 2 12:48:20.671246 systemd[1]: Created slice user.slice - User and Session Slice. Mar 2 12:48:20.671252 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 12:48:20.671258 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 12:48:20.671264 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 2 12:48:20.671270 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 2 12:48:20.671276 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 2 12:48:20.671283 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 12:48:20.671289 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Mar 2 12:48:20.671296 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 12:48:20.671302 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 12:48:20.671309 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 2 12:48:20.671315 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 2 12:48:20.671320 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 2 12:48:20.671326 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 2 12:48:20.671333 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 12:48:20.671339 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 12:48:20.671345 systemd[1]: Reached target slices.target - Slice Units. Mar 2 12:48:20.671351 systemd[1]: Reached target swap.target - Swaps. Mar 2 12:48:20.671357 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 2 12:48:20.671363 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 2 12:48:20.671371 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 2 12:48:20.671377 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 12:48:20.671384 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 12:48:20.671390 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 12:48:20.671396 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 2 12:48:20.671402 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 2 12:48:20.671408 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 2 12:48:20.671415 systemd[1]: Mounting media.mount - External Media Directory... Mar 2 12:48:20.671421 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 2 12:48:20.671427 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 2 12:48:20.671433 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 2 12:48:20.671440 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 2 12:48:20.671446 systemd[1]: Reached target machines.target - Containers. Mar 2 12:48:20.671452 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 2 12:48:20.671458 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 12:48:20.671465 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 12:48:20.671472 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 2 12:48:20.671478 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 12:48:20.671484 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 12:48:20.671499 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 12:48:20.671506 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 2 12:48:20.671512 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 12:48:20.671519 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 2 12:48:20.671525 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 2 12:48:20.671532 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 2 12:48:20.671539 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 2 12:48:20.671545 systemd[1]: Stopped systemd-fsck-usr.service. Mar 2 12:48:20.671551 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 12:48:20.671557 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 12:48:20.671563 kernel: fuse: init (API version 7.41) Mar 2 12:48:20.671569 kernel: loop: module loaded Mar 2 12:48:20.671585 systemd-journald[1391]: Collecting audit messages is disabled. Mar 2 12:48:20.671599 kernel: ACPI: bus type drm_connector registered Mar 2 12:48:20.671606 systemd-journald[1391]: Journal started Mar 2 12:48:20.671621 systemd-journald[1391]: Runtime Journal (/run/log/journal/d4cfac40b4944b17a40a00661ac4e34b) is 8M, max 78.3M, 70.3M free. Mar 2 12:48:20.008554 systemd[1]: Queued start job for default target multi-user.target. Mar 2 12:48:20.015882 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Mar 2 12:48:20.016232 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 2 12:48:20.016474 systemd[1]: systemd-journald.service: Consumed 2.372s CPU time. Mar 2 12:48:20.683415 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 12:48:20.696852 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 12:48:20.707175 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 2 12:48:20.719510 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 2 12:48:20.730700 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 12:48:20.734503 systemd[1]: verity-setup.service: Deactivated successfully. Mar 2 12:48:20.741258 systemd[1]: Stopped verity-setup.service. Mar 2 12:48:20.753265 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 12:48:20.753879 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 2 12:48:20.758191 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 2 12:48:20.763170 systemd[1]: Mounted media.mount - External Media Directory. Mar 2 12:48:20.767093 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 2 12:48:20.771729 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 2 12:48:20.776237 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 2 12:48:20.782521 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 2 12:48:20.789465 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 12:48:20.794817 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 2 12:48:20.794936 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 2 12:48:20.801902 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 12:48:20.802019 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 12:48:20.808691 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 12:48:20.808804 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 12:48:20.813283 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 12:48:20.813398 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 12:48:20.818695 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 2 12:48:20.818798 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 2 12:48:20.823389 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 12:48:20.823520 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 12:48:20.830588 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 12:48:20.835350 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 12:48:20.840424 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 2 12:48:20.846562 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 2 12:48:20.852032 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 12:48:20.863745 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 12:48:20.869036 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 2 12:48:20.882022 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 2 12:48:20.886545 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 2 12:48:20.886573 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 12:48:20.891206 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 2 12:48:20.907618 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 2 12:48:20.911793 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 12:48:20.926381 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 2 12:48:20.937118 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 2 12:48:20.941813 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 12:48:20.942653 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 2 12:48:20.947507 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 12:48:20.949610 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 12:48:20.954640 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 2 12:48:20.960299 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 12:48:20.967046 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 2 12:48:20.972125 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 2 12:48:20.977831 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 2 12:48:20.983785 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 2 12:48:20.989284 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 2 12:48:21.019752 systemd-journald[1391]: Time spent on flushing to /var/log/journal/d4cfac40b4944b17a40a00661ac4e34b is 42.109ms for 937 entries. Mar 2 12:48:21.019752 systemd-journald[1391]: System Journal (/var/log/journal/d4cfac40b4944b17a40a00661ac4e34b) is 11.8M, max 2.6G, 2.6G free. Mar 2 12:48:21.114433 systemd-journald[1391]: Received client request to flush runtime journal. Mar 2 12:48:21.114479 systemd-journald[1391]: /var/log/journal/d4cfac40b4944b17a40a00661ac4e34b/system.journal: Realtime clock jumped backwards relative to last journal entry, rotating. Mar 2 12:48:21.114516 systemd-journald[1391]: Rotating system journal. Mar 2 12:48:21.114535 kernel: loop0: detected capacity change from 0 to 27936 Mar 2 12:48:21.075509 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 2 12:48:21.076778 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 2 12:48:21.116528 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 2 12:48:21.126434 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:48:21.194920 systemd-tmpfiles[1442]: ACLs are not supported, ignoring. Mar 2 12:48:21.194931 systemd-tmpfiles[1442]: ACLs are not supported, ignoring. Mar 2 12:48:21.197705 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 12:48:21.204287 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 2 12:48:21.322830 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 2 12:48:21.328256 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 12:48:21.346733 systemd-tmpfiles[1460]: ACLs are not supported, ignoring. Mar 2 12:48:21.346938 systemd-tmpfiles[1460]: ACLs are not supported, ignoring. Mar 2 12:48:21.349359 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 12:48:21.412512 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 2 12:48:21.456299 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 2 12:48:21.462397 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 12:48:21.474517 kernel: loop1: detected capacity change from 0 to 197488 Mar 2 12:48:21.491296 systemd-udevd[1466]: Using default interface naming scheme 'v255'. Mar 2 12:48:21.558521 kernel: loop2: detected capacity change from 0 to 100632 Mar 2 12:48:21.668508 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 12:48:21.677846 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 12:48:21.724997 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Mar 2 12:48:21.743602 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 2 12:48:21.813236 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 2 12:48:21.818518 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#143 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 2 12:48:21.841520 kernel: hv_vmbus: registering driver hv_balloon Mar 2 12:48:21.841571 kernel: mousedev: PS/2 mouse device common for all mice Mar 2 12:48:21.849592 kernel: hv_vmbus: registering driver hyperv_fb Mar 2 12:48:21.854901 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Mar 2 12:48:21.854953 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Mar 2 12:48:21.860410 kernel: hv_balloon: Memory hot add disabled on ARM64 Mar 2 12:48:21.860467 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Mar 2 12:48:21.876791 kernel: Console: switching to colour dummy device 80x25 Mar 2 12:48:21.883858 kernel: Console: switching to colour frame buffer device 128x48 Mar 2 12:48:21.946372 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:48:21.957045 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:48:21.957174 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:21.963896 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:48:21.972451 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:48:21.974606 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:21.982645 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:48:22.015509 kernel: MACsec IEEE 802.1AE Mar 2 12:48:22.028515 kernel: loop3: detected capacity change from 0 to 119840 Mar 2 12:48:22.100860 systemd-networkd[1485]: lo: Link UP Mar 2 12:48:22.100866 systemd-networkd[1485]: lo: Gained carrier Mar 2 12:48:22.101749 systemd-networkd[1485]: Enumeration completed Mar 2 12:48:22.102565 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 12:48:22.102844 systemd-networkd[1485]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:48:22.102903 systemd-networkd[1485]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 12:48:22.109196 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Mar 2 12:48:22.115458 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 2 12:48:22.127605 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 2 12:48:22.133639 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 2 12:48:22.165505 kernel: mlx5_core 6b1f:00:02.0 enP27423s1: Link up Mar 2 12:48:22.186503 kernel: hv_netvsc 7ced8db6-fce0-7ced-8db6-fce07ced8db6 eth0: Data path switched to VF: enP27423s1 Mar 2 12:48:22.186851 systemd-networkd[1485]: enP27423s1: Link UP Mar 2 12:48:22.186972 systemd-networkd[1485]: eth0: Link UP Mar 2 12:48:22.186975 systemd-networkd[1485]: eth0: Gained carrier Mar 2 12:48:22.186987 systemd-networkd[1485]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:48:22.190856 systemd-networkd[1485]: enP27423s1: Gained carrier Mar 2 12:48:22.192801 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 2 12:48:22.199057 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 2 12:48:22.206595 systemd-networkd[1485]: eth0: DHCPv4 address 10.200.20.38/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 2 12:48:22.349754 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:48:22.375510 kernel: loop4: detected capacity change from 0 to 27936 Mar 2 12:48:22.387519 kernel: loop5: detected capacity change from 0 to 197488 Mar 2 12:48:22.407526 kernel: loop6: detected capacity change from 0 to 100632 Mar 2 12:48:22.418505 kernel: loop7: detected capacity change from 0 to 119840 Mar 2 12:48:22.426070 (sd-merge)[1616]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Mar 2 12:48:22.426416 (sd-merge)[1616]: Merged extensions into '/usr'. Mar 2 12:48:22.429963 systemd[1]: Reload requested from client PID 1440 ('systemd-sysext') (unit systemd-sysext.service)... Mar 2 12:48:22.429978 systemd[1]: Reloading... Mar 2 12:48:22.474561 zram_generator::config[1644]: No configuration found. Mar 2 12:48:22.636400 systemd[1]: Reloading finished in 206 ms. Mar 2 12:48:22.655521 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 2 12:48:22.667343 systemd[1]: Starting ensure-sysext.service... Mar 2 12:48:22.672598 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 12:48:22.688596 systemd[1]: Reload requested from client PID 1700 ('systemctl') (unit ensure-sysext.service)... Mar 2 12:48:22.688607 systemd[1]: Reloading... Mar 2 12:48:22.691995 systemd-tmpfiles[1701]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Mar 2 12:48:22.692255 systemd-tmpfiles[1701]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Mar 2 12:48:22.692541 systemd-tmpfiles[1701]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 2 12:48:22.692779 systemd-tmpfiles[1701]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 2 12:48:22.693298 systemd-tmpfiles[1701]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 2 12:48:22.693557 systemd-tmpfiles[1701]: ACLs are not supported, ignoring. Mar 2 12:48:22.693672 systemd-tmpfiles[1701]: ACLs are not supported, ignoring. Mar 2 12:48:22.696827 systemd-tmpfiles[1701]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 12:48:22.696964 systemd-tmpfiles[1701]: Skipping /boot Mar 2 12:48:22.702897 systemd-tmpfiles[1701]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 12:48:22.702980 systemd-tmpfiles[1701]: Skipping /boot Mar 2 12:48:22.741523 zram_generator::config[1735]: No configuration found. Mar 2 12:48:22.886888 systemd[1]: Reloading finished in 198 ms. Mar 2 12:48:22.903749 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 12:48:22.921608 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 2 12:48:22.930086 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 2 12:48:22.938402 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 2 12:48:22.944322 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 12:48:22.952658 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 2 12:48:22.960359 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 12:48:22.969426 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 12:48:22.976656 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 12:48:22.983640 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 12:48:22.987767 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 12:48:22.987849 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 12:48:22.990580 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 12:48:22.990708 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 12:48:22.995704 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 12:48:22.995819 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 12:48:23.001174 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 12:48:23.001287 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 12:48:23.011083 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 2 12:48:23.018955 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 12:48:23.020681 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 12:48:23.027591 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 12:48:23.036437 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 12:48:23.043619 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 12:48:23.043714 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 12:48:23.044526 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 2 12:48:23.050194 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 12:48:23.050308 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 12:48:23.055029 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 12:48:23.055136 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 12:48:23.061487 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 12:48:23.061619 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 12:48:23.070305 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 12:48:23.071198 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 12:48:23.076921 systemd-resolved[1792]: Positive Trust Anchors: Mar 2 12:48:23.077120 systemd-resolved[1792]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 12:48:23.077186 systemd-resolved[1792]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 12:48:23.081679 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 12:48:23.087674 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 12:48:23.094891 systemd-resolved[1792]: Using system hostname 'ci-4459.2.101-47783670b7'. Mar 2 12:48:23.097126 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 12:48:23.101645 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 12:48:23.101735 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 12:48:23.101837 systemd[1]: Reached target time-set.target - System Time Set. Mar 2 12:48:23.106874 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 12:48:23.111625 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 12:48:23.111765 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 12:48:23.117846 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 12:48:23.118200 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 12:48:23.123768 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 12:48:23.123901 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 12:48:23.129660 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 12:48:23.130083 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 12:48:23.139566 systemd[1]: Finished ensure-sysext.service. Mar 2 12:48:23.145109 systemd[1]: Reached target network.target - Network. Mar 2 12:48:23.148897 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 12:48:23.154217 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 12:48:23.154362 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 12:48:23.171308 augenrules[1837]: No rules Mar 2 12:48:23.172250 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 12:48:23.173529 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 2 12:48:23.282632 systemd-networkd[1485]: eth0: Gained IPv6LL Mar 2 12:48:23.284464 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 2 12:48:23.289808 systemd[1]: Reached target network-online.target - Network is Online. Mar 2 12:48:23.613528 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 2 12:48:23.618741 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 2 12:48:26.486126 ldconfig[1435]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 2 12:48:26.497554 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 2 12:48:26.506109 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 2 12:48:26.521395 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 2 12:48:26.526252 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 12:48:26.530665 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 2 12:48:26.535905 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 2 12:48:26.541021 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 2 12:48:26.545711 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 2 12:48:26.550892 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 2 12:48:26.556289 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 2 12:48:26.556383 systemd[1]: Reached target paths.target - Path Units. Mar 2 12:48:26.560250 systemd[1]: Reached target timers.target - Timer Units. Mar 2 12:48:26.565292 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 2 12:48:26.570971 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 2 12:48:26.576573 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 2 12:48:26.582171 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 2 12:48:26.587464 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 2 12:48:26.593252 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 2 12:48:26.597697 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 2 12:48:26.602787 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 2 12:48:26.607172 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 12:48:26.611101 systemd[1]: Reached target basic.target - Basic System. Mar 2 12:48:26.615050 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 2 12:48:26.615144 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 2 12:48:26.616978 systemd[1]: Starting chronyd.service - NTP client/server... Mar 2 12:48:26.628960 systemd[1]: Starting containerd.service - containerd container runtime... Mar 2 12:48:26.636353 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 2 12:48:26.644641 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 2 12:48:26.651484 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 2 12:48:26.658587 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 2 12:48:26.664522 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 2 12:48:26.669928 jq[1858]: false Mar 2 12:48:26.670314 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 2 12:48:26.672608 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Mar 2 12:48:26.676960 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Mar 2 12:48:26.678620 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:48:26.684641 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 2 12:48:26.689451 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 2 12:48:26.694569 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 2 12:48:26.701594 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 2 12:48:26.707034 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 2 12:48:26.714289 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 2 12:48:26.718370 extend-filesystems[1859]: Found /dev/sda6 Mar 2 12:48:26.721237 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 2 12:48:26.721569 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 2 12:48:26.724706 systemd[1]: Starting update-engine.service - Update Engine... Mar 2 12:48:26.734867 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 2 12:48:26.741908 extend-filesystems[1859]: Found /dev/sda9 Mar 2 12:48:26.746581 extend-filesystems[1859]: Checking size of /dev/sda9 Mar 2 12:48:26.744849 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 2 12:48:26.754148 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 2 12:48:26.772275 jq[1878]: true Mar 2 12:48:26.754562 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 2 12:48:26.756126 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 2 12:48:26.761826 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 2 12:48:26.776168 systemd[1]: motdgen.service: Deactivated successfully. Mar 2 12:48:26.776320 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 2 12:48:26.788781 update_engine[1875]: I20260302 12:48:26.788713 1875 main.cc:92] Flatcar Update Engine starting Mar 2 12:48:26.791726 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 2 12:48:26.805639 extend-filesystems[1859]: Old size kept for /dev/sda9 Mar 2 12:48:26.819983 jq[1895]: true Mar 2 12:48:26.806348 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 2 12:48:26.807903 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 2 12:48:26.812648 (ntainerd)[1899]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 2 12:48:26.856047 systemd-logind[1870]: New seat seat0. Mar 2 12:48:26.857874 systemd-logind[1870]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 2 12:48:26.858011 systemd[1]: Started systemd-logind.service - User Login Management. Mar 2 12:48:26.867525 tar[1889]: linux-arm64/LICENSE Mar 2 12:48:26.868343 tar[1889]: linux-arm64/helm Mar 2 12:48:26.914972 KVP[1860]: KVP starting; pid is:1860 Mar 2 12:48:26.919522 chronyd[1850]: chronyd version 4.7 starting (+CMDMON +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +NTS +SECHASH +IPV6 -DEBUG) Mar 2 12:48:26.923904 KVP[1860]: KVP LIC Version: 3.1 Mar 2 12:48:26.924515 kernel: hv_utils: KVP IC version 4.0 Mar 2 12:48:26.937871 bash[1938]: Updated "/home/core/.ssh/authorized_keys" Mar 2 12:48:26.940900 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 2 12:48:26.950167 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 2 12:48:26.983287 chronyd[1850]: Timezone right/UTC failed leap second check, ignoring Mar 2 12:48:26.983560 systemd[1]: Started chronyd.service - NTP client/server. Mar 2 12:48:26.983429 chronyd[1850]: Loaded seccomp filter (level 2) Mar 2 12:48:27.010195 dbus-daemon[1853]: [system] SELinux support is enabled Mar 2 12:48:27.010311 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 2 12:48:27.022873 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 2 12:48:27.022896 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 2 12:48:27.030500 update_engine[1875]: I20260302 12:48:27.030359 1875 update_check_scheduler.cc:74] Next update check in 11m8s Mar 2 12:48:27.033650 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 2 12:48:27.033667 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 2 12:48:27.048008 dbus-daemon[1853]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 2 12:48:27.053550 systemd[1]: Started update-engine.service - Update Engine. Mar 2 12:48:27.065024 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 2 12:48:27.110121 coreos-metadata[1852]: Mar 02 12:48:27.109 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 2 12:48:27.112808 coreos-metadata[1852]: Mar 02 12:48:27.112 INFO Fetch successful Mar 2 12:48:27.112808 coreos-metadata[1852]: Mar 02 12:48:27.112 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Mar 2 12:48:27.117116 coreos-metadata[1852]: Mar 02 12:48:27.116 INFO Fetch successful Mar 2 12:48:27.117116 coreos-metadata[1852]: Mar 02 12:48:27.117 INFO Fetching http://168.63.129.16/machine/633558f2-d135-4acc-91cc-628a95a390d6/468321a7%2D5aa9%2D48fb%2D9ce2%2D80c309b62095.%5Fci%2D4459.2.101%2D47783670b7?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Mar 2 12:48:27.119234 coreos-metadata[1852]: Mar 02 12:48:27.119 INFO Fetch successful Mar 2 12:48:27.119450 coreos-metadata[1852]: Mar 02 12:48:27.119 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Mar 2 12:48:27.127994 coreos-metadata[1852]: Mar 02 12:48:27.127 INFO Fetch successful Mar 2 12:48:27.148928 sshd_keygen[1891]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 2 12:48:27.160311 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 2 12:48:27.167009 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 2 12:48:27.180398 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 2 12:48:27.189365 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 2 12:48:27.202690 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Mar 2 12:48:27.220786 systemd[1]: issuegen.service: Deactivated successfully. Mar 2 12:48:27.221865 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 2 12:48:27.230400 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 2 12:48:27.253288 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 2 12:48:27.260607 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Mar 2 12:48:27.272124 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 2 12:48:27.278994 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Mar 2 12:48:27.283889 systemd[1]: Reached target getty.target - Login Prompts. Mar 2 12:48:27.349756 tar[1889]: linux-arm64/README.md Mar 2 12:48:27.362672 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 2 12:48:27.372617 locksmithd[1997]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 2 12:48:27.537410 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:48:27.576205 containerd[1899]: time="2026-03-02T12:48:27Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Mar 2 12:48:27.577009 containerd[1899]: time="2026-03-02T12:48:27.576974240Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Mar 2 12:48:27.582393 containerd[1899]: time="2026-03-02T12:48:27.582363736Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="6.752µs" Mar 2 12:48:27.582393 containerd[1899]: time="2026-03-02T12:48:27.582388536Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Mar 2 12:48:27.582477 containerd[1899]: time="2026-03-02T12:48:27.582402816Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Mar 2 12:48:27.582572 containerd[1899]: time="2026-03-02T12:48:27.582552696Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Mar 2 12:48:27.582572 containerd[1899]: time="2026-03-02T12:48:27.582570352Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Mar 2 12:48:27.582604 containerd[1899]: time="2026-03-02T12:48:27.582586912Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 12:48:27.582642 containerd[1899]: time="2026-03-02T12:48:27.582627112Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 12:48:27.582642 containerd[1899]: time="2026-03-02T12:48:27.582638384Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 12:48:27.582798 containerd[1899]: time="2026-03-02T12:48:27.582781648Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 12:48:27.582798 containerd[1899]: time="2026-03-02T12:48:27.582796472Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 12:48:27.582826 containerd[1899]: time="2026-03-02T12:48:27.582808632Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 12:48:27.582826 containerd[1899]: time="2026-03-02T12:48:27.582813584Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Mar 2 12:48:27.582895 containerd[1899]: time="2026-03-02T12:48:27.582882928Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Mar 2 12:48:27.583035 containerd[1899]: time="2026-03-02T12:48:27.583019664Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 12:48:27.583057 containerd[1899]: time="2026-03-02T12:48:27.583044136Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 12:48:27.583071 containerd[1899]: time="2026-03-02T12:48:27.583056576Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Mar 2 12:48:27.583098 containerd[1899]: time="2026-03-02T12:48:27.583083464Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Mar 2 12:48:27.583246 containerd[1899]: time="2026-03-02T12:48:27.583220024Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Mar 2 12:48:27.583304 containerd[1899]: time="2026-03-02T12:48:27.583290488Z" level=info msg="metadata content store policy set" policy=shared Mar 2 12:48:27.599074 containerd[1899]: time="2026-03-02T12:48:27.599049024Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Mar 2 12:48:27.599130 containerd[1899]: time="2026-03-02T12:48:27.599089544Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Mar 2 12:48:27.599130 containerd[1899]: time="2026-03-02T12:48:27.599100176Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Mar 2 12:48:27.599130 containerd[1899]: time="2026-03-02T12:48:27.599108216Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Mar 2 12:48:27.599130 containerd[1899]: time="2026-03-02T12:48:27.599116584Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Mar 2 12:48:27.599130 containerd[1899]: time="2026-03-02T12:48:27.599122896Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Mar 2 12:48:27.599198 containerd[1899]: time="2026-03-02T12:48:27.599130848Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Mar 2 12:48:27.599198 containerd[1899]: time="2026-03-02T12:48:27.599145272Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Mar 2 12:48:27.599198 containerd[1899]: time="2026-03-02T12:48:27.599152536Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Mar 2 12:48:27.599198 containerd[1899]: time="2026-03-02T12:48:27.599158784Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Mar 2 12:48:27.599198 containerd[1899]: time="2026-03-02T12:48:27.599164584Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Mar 2 12:48:27.599198 containerd[1899]: time="2026-03-02T12:48:27.599175552Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Mar 2 12:48:27.599282 containerd[1899]: time="2026-03-02T12:48:27.599267608Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Mar 2 12:48:27.599306 containerd[1899]: time="2026-03-02T12:48:27.599286272Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Mar 2 12:48:27.599306 containerd[1899]: time="2026-03-02T12:48:27.599296128Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Mar 2 12:48:27.599306 containerd[1899]: time="2026-03-02T12:48:27.599303480Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Mar 2 12:48:27.599340 containerd[1899]: time="2026-03-02T12:48:27.599310832Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Mar 2 12:48:27.599340 containerd[1899]: time="2026-03-02T12:48:27.599317824Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Mar 2 12:48:27.599340 containerd[1899]: time="2026-03-02T12:48:27.599330160Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Mar 2 12:48:27.599340 containerd[1899]: time="2026-03-02T12:48:27.599337072Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Mar 2 12:48:27.599405 containerd[1899]: time="2026-03-02T12:48:27.599343680Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Mar 2 12:48:27.599405 containerd[1899]: time="2026-03-02T12:48:27.599350232Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Mar 2 12:48:27.599405 containerd[1899]: time="2026-03-02T12:48:27.599356424Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Mar 2 12:48:27.599405 containerd[1899]: time="2026-03-02T12:48:27.599391712Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Mar 2 12:48:27.599405 containerd[1899]: time="2026-03-02T12:48:27.599400536Z" level=info msg="Start snapshots syncer" Mar 2 12:48:27.599479 containerd[1899]: time="2026-03-02T12:48:27.599412752Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Mar 2 12:48:27.599852 containerd[1899]: time="2026-03-02T12:48:27.599818376Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Mar 2 12:48:27.599946 containerd[1899]: time="2026-03-02T12:48:27.599874416Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Mar 2 12:48:27.599946 containerd[1899]: time="2026-03-02T12:48:27.599915680Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Mar 2 12:48:27.600040 containerd[1899]: time="2026-03-02T12:48:27.600016368Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Mar 2 12:48:27.600067 containerd[1899]: time="2026-03-02T12:48:27.600041656Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Mar 2 12:48:27.600067 containerd[1899]: time="2026-03-02T12:48:27.600052712Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Mar 2 12:48:27.600067 containerd[1899]: time="2026-03-02T12:48:27.600063392Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Mar 2 12:48:27.600109 containerd[1899]: time="2026-03-02T12:48:27.600074000Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Mar 2 12:48:27.600109 containerd[1899]: time="2026-03-02T12:48:27.600081784Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Mar 2 12:48:27.600109 containerd[1899]: time="2026-03-02T12:48:27.600091328Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Mar 2 12:48:27.600157 containerd[1899]: time="2026-03-02T12:48:27.600112456Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Mar 2 12:48:27.600157 containerd[1899]: time="2026-03-02T12:48:27.600123600Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Mar 2 12:48:27.600157 containerd[1899]: time="2026-03-02T12:48:27.600133128Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Mar 2 12:48:27.600198 containerd[1899]: time="2026-03-02T12:48:27.600160720Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 12:48:27.600198 containerd[1899]: time="2026-03-02T12:48:27.600170616Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 12:48:27.600198 containerd[1899]: time="2026-03-02T12:48:27.600178528Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 12:48:27.600198 containerd[1899]: time="2026-03-02T12:48:27.600186656Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 12:48:27.600198 containerd[1899]: time="2026-03-02T12:48:27.600193992Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Mar 2 12:48:27.600198 containerd[1899]: time="2026-03-02T12:48:27.600200376Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Mar 2 12:48:27.600299 containerd[1899]: time="2026-03-02T12:48:27.600209832Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Mar 2 12:48:27.600299 containerd[1899]: time="2026-03-02T12:48:27.600224120Z" level=info msg="runtime interface created" Mar 2 12:48:27.600299 containerd[1899]: time="2026-03-02T12:48:27.600228064Z" level=info msg="created NRI interface" Mar 2 12:48:27.600299 containerd[1899]: time="2026-03-02T12:48:27.600237112Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Mar 2 12:48:27.600299 containerd[1899]: time="2026-03-02T12:48:27.600247992Z" level=info msg="Connect containerd service" Mar 2 12:48:27.600370 containerd[1899]: time="2026-03-02T12:48:27.600301208Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 2 12:48:27.601622 containerd[1899]: time="2026-03-02T12:48:27.601592200Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 12:48:27.677799 (kubelet)[2043]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:48:27.954256 kubelet[2043]: E0302 12:48:27.954189 2043 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:48:27.956278 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:48:27.956387 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:48:27.956741 systemd[1]: kubelet.service: Consumed 485ms CPU time, 244.3M memory peak. Mar 2 12:48:28.107903 containerd[1899]: time="2026-03-02T12:48:28.107845976Z" level=info msg="Start subscribing containerd event" Mar 2 12:48:28.107987 containerd[1899]: time="2026-03-02T12:48:28.107916848Z" level=info msg="Start recovering state" Mar 2 12:48:28.108011 containerd[1899]: time="2026-03-02T12:48:28.107990880Z" level=info msg="Start event monitor" Mar 2 12:48:28.108011 containerd[1899]: time="2026-03-02T12:48:28.108002872Z" level=info msg="Start cni network conf syncer for default" Mar 2 12:48:28.108011 containerd[1899]: time="2026-03-02T12:48:28.108007456Z" level=info msg="Start streaming server" Mar 2 12:48:28.108050 containerd[1899]: time="2026-03-02T12:48:28.108014736Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Mar 2 12:48:28.108050 containerd[1899]: time="2026-03-02T12:48:28.108020472Z" level=info msg="runtime interface starting up..." Mar 2 12:48:28.108050 containerd[1899]: time="2026-03-02T12:48:28.108024256Z" level=info msg="starting plugins..." Mar 2 12:48:28.108050 containerd[1899]: time="2026-03-02T12:48:28.108033968Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Mar 2 12:48:28.108398 containerd[1899]: time="2026-03-02T12:48:28.108375224Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 2 12:48:28.108510 containerd[1899]: time="2026-03-02T12:48:28.108416280Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 2 12:48:28.108510 containerd[1899]: time="2026-03-02T12:48:28.108458136Z" level=info msg="containerd successfully booted in 0.532678s" Mar 2 12:48:28.108568 systemd[1]: Started containerd.service - containerd container runtime. Mar 2 12:48:28.114047 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 2 12:48:28.120566 systemd[1]: Startup finished in 1.684s (kernel) + 11.927s (initrd) + 11.032s (userspace) = 24.645s. Mar 2 12:48:28.444878 login[2026]: pam_lastlog(login:session): file /var/log/lastlog is locked/write, retrying Mar 2 12:48:28.445645 login[2025]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:28.451753 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 2 12:48:28.455684 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 2 12:48:28.460589 systemd-logind[1870]: New session 1 of user core. Mar 2 12:48:28.499889 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 2 12:48:28.502034 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 2 12:48:28.545936 (systemd)[2071]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 2 12:48:28.547872 systemd-logind[1870]: New session c1 of user core. Mar 2 12:48:28.731169 systemd[2071]: Queued start job for default target default.target. Mar 2 12:48:28.740281 systemd[2071]: Created slice app.slice - User Application Slice. Mar 2 12:48:28.740380 systemd[2071]: Reached target paths.target - Paths. Mar 2 12:48:28.740478 systemd[2071]: Reached target timers.target - Timers. Mar 2 12:48:28.741520 systemd[2071]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 2 12:48:28.750030 systemd[2071]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 2 12:48:28.750074 systemd[2071]: Reached target sockets.target - Sockets. Mar 2 12:48:28.750112 systemd[2071]: Reached target basic.target - Basic System. Mar 2 12:48:28.750133 systemd[2071]: Reached target default.target - Main User Target. Mar 2 12:48:28.750152 systemd[2071]: Startup finished in 197ms. Mar 2 12:48:28.750216 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 2 12:48:28.758905 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 2 12:48:29.052181 waagent[2023]: 2026-03-02T12:48:29.052075Z INFO Daemon Daemon Azure Linux Agent Version: 2.12.0.4 Mar 2 12:48:29.060194 waagent[2023]: 2026-03-02T12:48:29.056567Z INFO Daemon Daemon OS: flatcar 4459.2.101 Mar 2 12:48:29.060565 waagent[2023]: 2026-03-02T12:48:29.060527Z INFO Daemon Daemon Python: 3.11.13 Mar 2 12:48:29.064430 waagent[2023]: 2026-03-02T12:48:29.064389Z INFO Daemon Daemon Run daemon Mar 2 12:48:29.067774 waagent[2023]: 2026-03-02T12:48:29.067738Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4459.2.101' Mar 2 12:48:29.074508 waagent[2023]: 2026-03-02T12:48:29.074467Z INFO Daemon Daemon Using waagent for provisioning Mar 2 12:48:29.078358 waagent[2023]: 2026-03-02T12:48:29.078324Z INFO Daemon Daemon Activate resource disk Mar 2 12:48:29.081917 waagent[2023]: 2026-03-02T12:48:29.081883Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Mar 2 12:48:29.089852 waagent[2023]: 2026-03-02T12:48:29.089815Z INFO Daemon Daemon Found device: None Mar 2 12:48:29.093063 waagent[2023]: 2026-03-02T12:48:29.093030Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Mar 2 12:48:29.099309 waagent[2023]: 2026-03-02T12:48:29.099276Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Mar 2 12:48:29.107677 waagent[2023]: 2026-03-02T12:48:29.107640Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 2 12:48:29.112050 waagent[2023]: 2026-03-02T12:48:29.112014Z INFO Daemon Daemon Running default provisioning handler Mar 2 12:48:29.120141 waagent[2023]: 2026-03-02T12:48:29.120109Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Mar 2 12:48:29.129787 waagent[2023]: 2026-03-02T12:48:29.129756Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Mar 2 12:48:29.136584 waagent[2023]: 2026-03-02T12:48:29.136554Z INFO Daemon Daemon cloud-init is enabled: False Mar 2 12:48:29.140176 waagent[2023]: 2026-03-02T12:48:29.140146Z INFO Daemon Daemon Copying ovf-env.xml Mar 2 12:48:29.200005 waagent[2023]: 2026-03-02T12:48:29.199969Z INFO Daemon Daemon Successfully mounted dvd Mar 2 12:48:29.226878 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Mar 2 12:48:29.228718 waagent[2023]: 2026-03-02T12:48:29.228673Z INFO Daemon Daemon Detect protocol endpoint Mar 2 12:48:29.232120 waagent[2023]: 2026-03-02T12:48:29.232087Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 2 12:48:29.236139 waagent[2023]: 2026-03-02T12:48:29.236111Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Mar 2 12:48:29.240884 waagent[2023]: 2026-03-02T12:48:29.240860Z INFO Daemon Daemon Test for route to 168.63.129.16 Mar 2 12:48:29.244610 waagent[2023]: 2026-03-02T12:48:29.244578Z INFO Daemon Daemon Route to 168.63.129.16 exists Mar 2 12:48:29.248192 waagent[2023]: 2026-03-02T12:48:29.248163Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Mar 2 12:48:29.288973 waagent[2023]: 2026-03-02T12:48:29.288940Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Mar 2 12:48:29.293769 waagent[2023]: 2026-03-02T12:48:29.293748Z INFO Daemon Daemon Wire protocol version:2012-11-30 Mar 2 12:48:29.297542 waagent[2023]: 2026-03-02T12:48:29.297517Z INFO Daemon Daemon Server preferred version:2015-04-05 Mar 2 12:48:29.438621 waagent[2023]: 2026-03-02T12:48:29.438560Z INFO Daemon Daemon Initializing goal state during protocol detection Mar 2 12:48:29.443162 waagent[2023]: 2026-03-02T12:48:29.443128Z INFO Daemon Daemon Forcing an update of the goal state. Mar 2 12:48:29.445976 login[2026]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:29.450972 waagent[2023]: 2026-03-02T12:48:29.450931Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Mar 2 12:48:29.457399 systemd-logind[1870]: New session 2 of user core. Mar 2 12:48:29.462625 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 2 12:48:29.474375 waagent[2023]: 2026-03-02T12:48:29.474334Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.179 Mar 2 12:48:29.479561 waagent[2023]: 2026-03-02T12:48:29.479520Z INFO Daemon Mar 2 12:48:29.481855 waagent[2023]: 2026-03-02T12:48:29.481739Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: 82931bb2-45e2-4f1c-bdc9-e4705646fdb4 eTag: 2011553029784729335 source: Fabric] Mar 2 12:48:29.489908 waagent[2023]: 2026-03-02T12:48:29.489871Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Mar 2 12:48:29.495256 waagent[2023]: 2026-03-02T12:48:29.495219Z INFO Daemon Mar 2 12:48:29.497467 waagent[2023]: 2026-03-02T12:48:29.497441Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Mar 2 12:48:29.505739 waagent[2023]: 2026-03-02T12:48:29.505712Z INFO Daemon Daemon Downloading artifacts profile blob Mar 2 12:48:29.559107 waagent[2023]: 2026-03-02T12:48:29.559058Z INFO Daemon Downloaded certificate {'thumbprint': 'CC622D4EB889CFB95FBA39C1306EB6DA5ECA2D44', 'hasPrivateKey': True} Mar 2 12:48:29.566229 waagent[2023]: 2026-03-02T12:48:29.566194Z INFO Daemon Fetch goal state completed Mar 2 12:48:29.597503 waagent[2023]: 2026-03-02T12:48:29.597470Z INFO Daemon Daemon Starting provisioning Mar 2 12:48:29.600957 waagent[2023]: 2026-03-02T12:48:29.600919Z INFO Daemon Daemon Handle ovf-env.xml. Mar 2 12:48:29.604390 waagent[2023]: 2026-03-02T12:48:29.604363Z INFO Daemon Daemon Set hostname [ci-4459.2.101-47783670b7] Mar 2 12:48:29.626049 waagent[2023]: 2026-03-02T12:48:29.626010Z INFO Daemon Daemon Publish hostname [ci-4459.2.101-47783670b7] Mar 2 12:48:29.630675 waagent[2023]: 2026-03-02T12:48:29.630641Z INFO Daemon Daemon Examine /proc/net/route for primary interface Mar 2 12:48:29.635343 waagent[2023]: 2026-03-02T12:48:29.635312Z INFO Daemon Daemon Primary interface is [eth0] Mar 2 12:48:29.644385 systemd-networkd[1485]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:48:29.644390 systemd-networkd[1485]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 12:48:29.644412 systemd-networkd[1485]: eth0: DHCP lease lost Mar 2 12:48:29.645177 waagent[2023]: 2026-03-02T12:48:29.645042Z INFO Daemon Daemon Create user account if not exists Mar 2 12:48:29.649153 waagent[2023]: 2026-03-02T12:48:29.649120Z INFO Daemon Daemon User core already exists, skip useradd Mar 2 12:48:29.653443 waagent[2023]: 2026-03-02T12:48:29.653408Z INFO Daemon Daemon Configure sudoer Mar 2 12:48:29.660861 waagent[2023]: 2026-03-02T12:48:29.660821Z INFO Daemon Daemon Configure sshd Mar 2 12:48:29.667475 waagent[2023]: 2026-03-02T12:48:29.667439Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Mar 2 12:48:29.676472 waagent[2023]: 2026-03-02T12:48:29.676441Z INFO Daemon Daemon Deploy ssh public key. Mar 2 12:48:29.681537 systemd-networkd[1485]: eth0: DHCPv4 address 10.200.20.38/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 2 12:48:30.770604 waagent[2023]: 2026-03-02T12:48:30.770549Z INFO Daemon Daemon Provisioning complete Mar 2 12:48:30.784154 waagent[2023]: 2026-03-02T12:48:30.784121Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Mar 2 12:48:30.788530 waagent[2023]: 2026-03-02T12:48:30.788501Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Mar 2 12:48:30.795687 waagent[2023]: 2026-03-02T12:48:30.795661Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.12.0.4 is the most current agent Mar 2 12:48:30.895131 waagent[2121]: 2026-03-02T12:48:30.895076Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.12.0.4) Mar 2 12:48:30.896480 waagent[2121]: 2026-03-02T12:48:30.895463Z INFO ExtHandler ExtHandler OS: flatcar 4459.2.101 Mar 2 12:48:30.896480 waagent[2121]: 2026-03-02T12:48:30.895541Z INFO ExtHandler ExtHandler Python: 3.11.13 Mar 2 12:48:30.896480 waagent[2121]: 2026-03-02T12:48:30.895587Z INFO ExtHandler ExtHandler CPU Arch: aarch64 Mar 2 12:48:30.942356 waagent[2121]: 2026-03-02T12:48:30.942316Z INFO ExtHandler ExtHandler Distro: flatcar-4459.2.101; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.13; Arch: aarch64; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.22.0; Mar 2 12:48:30.942590 waagent[2121]: 2026-03-02T12:48:30.942558Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 2 12:48:30.942735 waagent[2121]: 2026-03-02T12:48:30.942709Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 2 12:48:30.948142 waagent[2121]: 2026-03-02T12:48:30.948096Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Mar 2 12:48:30.952683 waagent[2121]: 2026-03-02T12:48:30.952653Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.179 Mar 2 12:48:30.953111 waagent[2121]: 2026-03-02T12:48:30.953078Z INFO ExtHandler Mar 2 12:48:30.953235 waagent[2121]: 2026-03-02T12:48:30.953211Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 3dc68e21-3803-4008-a4fb-944f5ef42517 eTag: 2011553029784729335 source: Fabric] Mar 2 12:48:30.953564 waagent[2121]: 2026-03-02T12:48:30.953531Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Mar 2 12:48:30.954054 waagent[2121]: 2026-03-02T12:48:30.954023Z INFO ExtHandler Mar 2 12:48:30.954184 waagent[2121]: 2026-03-02T12:48:30.954160Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Mar 2 12:48:30.957407 waagent[2121]: 2026-03-02T12:48:30.957372Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Mar 2 12:48:31.012167 waagent[2121]: 2026-03-02T12:48:31.012130Z INFO ExtHandler Downloaded certificate {'thumbprint': 'CC622D4EB889CFB95FBA39C1306EB6DA5ECA2D44', 'hasPrivateKey': True} Mar 2 12:48:31.012620 waagent[2121]: 2026-03-02T12:48:31.012589Z INFO ExtHandler Fetch goal state completed Mar 2 12:48:31.023598 waagent[2121]: 2026-03-02T12:48:31.023535Z INFO ExtHandler ExtHandler OpenSSL version: OpenSSL 3.4.4 27 Jan 2026 (Library: OpenSSL 3.4.4 27 Jan 2026) Mar 2 12:48:31.026810 waagent[2121]: 2026-03-02T12:48:31.026776Z INFO ExtHandler ExtHandler WALinuxAgent-2.12.0.4 running as process 2121 Mar 2 12:48:31.027007 waagent[2121]: 2026-03-02T12:48:31.026978Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Mar 2 12:48:31.027312 waagent[2121]: 2026-03-02T12:48:31.027284Z INFO ExtHandler ExtHandler ******** AutoUpdate.UpdateToLatestVersion is set to False, not processing the operation ******** Mar 2 12:48:31.028512 waagent[2121]: 2026-03-02T12:48:31.028430Z INFO ExtHandler ExtHandler [CGI] Cgroup monitoring is not supported on ['flatcar', '4459.2.101', '', 'Flatcar Container Linux by Kinvolk'] Mar 2 12:48:31.028898 waagent[2121]: 2026-03-02T12:48:31.028866Z INFO ExtHandler ExtHandler [CGI] Agent will reset the quotas in case distro: ['flatcar', '4459.2.101', '', 'Flatcar Container Linux by Kinvolk'] went from supported to unsupported Mar 2 12:48:31.029101 waagent[2121]: 2026-03-02T12:48:31.029072Z INFO ExtHandler ExtHandler [CGI] Agent cgroups enabled: False Mar 2 12:48:31.029642 waagent[2121]: 2026-03-02T12:48:31.029612Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Mar 2 12:48:31.084344 waagent[2121]: 2026-03-02T12:48:31.084320Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Mar 2 12:48:31.084607 waagent[2121]: 2026-03-02T12:48:31.084574Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Mar 2 12:48:31.088811 waagent[2121]: 2026-03-02T12:48:31.088786Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Mar 2 12:48:31.093128 systemd[1]: Reload requested from client PID 2136 ('systemctl') (unit waagent.service)... Mar 2 12:48:31.093141 systemd[1]: Reloading... Mar 2 12:48:31.170512 zram_generator::config[2182]: No configuration found. Mar 2 12:48:31.298164 systemd[1]: Reloading finished in 204 ms. Mar 2 12:48:31.326520 waagent[2121]: 2026-03-02T12:48:31.324167Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Mar 2 12:48:31.326520 waagent[2121]: 2026-03-02T12:48:31.324279Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Mar 2 12:48:31.978134 waagent[2121]: 2026-03-02T12:48:31.977377Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Mar 2 12:48:31.978134 waagent[2121]: 2026-03-02T12:48:31.977700Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: 1. configuration enabled [True], 2. cgroups v1 enabled [False] OR cgroups v2 is in use and v2 resource limiting configuration enabled [False], 3. python supported: [True] Mar 2 12:48:31.978416 waagent[2121]: 2026-03-02T12:48:31.978328Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 2 12:48:31.978416 waagent[2121]: 2026-03-02T12:48:31.978390Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 2 12:48:31.978593 waagent[2121]: 2026-03-02T12:48:31.978558Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Mar 2 12:48:31.978709 waagent[2121]: 2026-03-02T12:48:31.978663Z INFO ExtHandler ExtHandler Starting env monitor service. Mar 2 12:48:31.978830 waagent[2121]: 2026-03-02T12:48:31.978799Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Mar 2 12:48:31.978830 waagent[2121]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Mar 2 12:48:31.978830 waagent[2121]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Mar 2 12:48:31.978830 waagent[2121]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Mar 2 12:48:31.978830 waagent[2121]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Mar 2 12:48:31.978830 waagent[2121]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 2 12:48:31.978830 waagent[2121]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 2 12:48:31.979334 waagent[2121]: 2026-03-02T12:48:31.979299Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Mar 2 12:48:31.979718 waagent[2121]: 2026-03-02T12:48:31.979684Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Mar 2 12:48:31.979760 waagent[2121]: 2026-03-02T12:48:31.979733Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 2 12:48:31.979827 waagent[2121]: 2026-03-02T12:48:31.979803Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Mar 2 12:48:31.979951 waagent[2121]: 2026-03-02T12:48:31.979918Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 2 12:48:31.980145 waagent[2121]: 2026-03-02T12:48:31.980112Z INFO EnvHandler ExtHandler Configure routes Mar 2 12:48:31.980205 waagent[2121]: 2026-03-02T12:48:31.980181Z INFO EnvHandler ExtHandler Gateway:None Mar 2 12:48:31.980234 waagent[2121]: 2026-03-02T12:48:31.980217Z INFO EnvHandler ExtHandler Routes:None Mar 2 12:48:31.980586 waagent[2121]: 2026-03-02T12:48:31.980552Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Mar 2 12:48:31.980797 waagent[2121]: 2026-03-02T12:48:31.980763Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Mar 2 12:48:31.980900 waagent[2121]: 2026-03-02T12:48:31.980873Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Mar 2 12:48:31.986126 waagent[2121]: 2026-03-02T12:48:31.986097Z INFO ExtHandler ExtHandler Mar 2 12:48:31.986265 waagent[2121]: 2026-03-02T12:48:31.986239Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: 1f758887-8c26-40b7-8cc8-9b89418bfda9 correlation 2bcb0171-5190-4f69-85c3-8f6bd569ee79 created: 2026-03-02T12:47:33.577827Z] Mar 2 12:48:31.986625 waagent[2121]: 2026-03-02T12:48:31.986594Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Mar 2 12:48:31.987089 waagent[2121]: 2026-03-02T12:48:31.987062Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 0 ms] Mar 2 12:48:32.016309 waagent[2121]: 2026-03-02T12:48:32.015477Z WARNING ExtHandler ExtHandler Failed to get firewall packets: 'iptables -w -t security -L OUTPUT --zero OUTPUT -nxv' failed: 2 (iptables v1.8.11 (nf_tables): Illegal option `--numeric' with this command Mar 2 12:48:32.016309 waagent[2121]: Try `iptables -h' or 'iptables --help' for more information.) Mar 2 12:48:32.016309 waagent[2121]: 2026-03-02T12:48:32.015789Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.12.0.4 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 1D3C399F-3C2F-428A-9373-5AF7A43BBD93;DroppedPackets: -1;UpdateGSErrors: 0;AutoUpdate: 0;UpdateMode: SelfUpdate;] Mar 2 12:48:32.038993 waagent[2121]: 2026-03-02T12:48:32.038943Z INFO MonitorHandler ExtHandler Network interfaces: Mar 2 12:48:32.038993 waagent[2121]: Executing ['ip', '-a', '-o', 'link']: Mar 2 12:48:32.038993 waagent[2121]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Mar 2 12:48:32.038993 waagent[2121]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:b6:fc:e0 brd ff:ff:ff:ff:ff:ff Mar 2 12:48:32.038993 waagent[2121]: 3: enP27423s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:b6:fc:e0 brd ff:ff:ff:ff:ff:ff\ altname enP27423p0s2 Mar 2 12:48:32.038993 waagent[2121]: Executing ['ip', '-4', '-a', '-o', 'address']: Mar 2 12:48:32.038993 waagent[2121]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Mar 2 12:48:32.038993 waagent[2121]: 2: eth0 inet 10.200.20.38/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Mar 2 12:48:32.038993 waagent[2121]: Executing ['ip', '-6', '-a', '-o', 'address']: Mar 2 12:48:32.038993 waagent[2121]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Mar 2 12:48:32.038993 waagent[2121]: 2: eth0 inet6 fe80::7eed:8dff:feb6:fce0/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Mar 2 12:48:32.102351 waagent[2121]: 2026-03-02T12:48:32.102306Z INFO EnvHandler ExtHandler Created firewall rules for the Azure Fabric: Mar 2 12:48:32.102351 waagent[2121]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:48:32.102351 waagent[2121]: pkts bytes target prot opt in out source destination Mar 2 12:48:32.102351 waagent[2121]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:48:32.102351 waagent[2121]: pkts bytes target prot opt in out source destination Mar 2 12:48:32.102351 waagent[2121]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:48:32.102351 waagent[2121]: pkts bytes target prot opt in out source destination Mar 2 12:48:32.102351 waagent[2121]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 2 12:48:32.102351 waagent[2121]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 2 12:48:32.102351 waagent[2121]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 2 12:48:32.104524 waagent[2121]: 2026-03-02T12:48:32.104475Z INFO EnvHandler ExtHandler Current Firewall rules: Mar 2 12:48:32.104524 waagent[2121]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:48:32.104524 waagent[2121]: pkts bytes target prot opt in out source destination Mar 2 12:48:32.104524 waagent[2121]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:48:32.104524 waagent[2121]: pkts bytes target prot opt in out source destination Mar 2 12:48:32.104524 waagent[2121]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:48:32.104524 waagent[2121]: pkts bytes target prot opt in out source destination Mar 2 12:48:32.104524 waagent[2121]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 2 12:48:32.104524 waagent[2121]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 2 12:48:32.104524 waagent[2121]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 2 12:48:32.104695 waagent[2121]: 2026-03-02T12:48:32.104675Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Mar 2 12:48:38.083006 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 2 12:48:38.084229 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:48:38.193501 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:48:38.196670 (kubelet)[2270]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:48:38.294759 kubelet[2270]: E0302 12:48:38.294700 2270 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:48:38.296933 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:48:38.297039 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:48:38.297443 systemd[1]: kubelet.service: Consumed 108ms CPU time, 105.7M memory peak. Mar 2 12:48:48.333168 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 2 12:48:48.334984 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:48:48.688367 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:48:48.690683 (kubelet)[2284]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:48:48.713546 kubelet[2284]: E0302 12:48:48.713509 2284 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:48:48.715486 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:48:48.715675 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:48:48.716082 systemd[1]: kubelet.service: Consumed 101ms CPU time, 105.1M memory peak. Mar 2 12:48:49.679354 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 2 12:48:49.681690 systemd[1]: Started sshd@0-10.200.20.38:22-10.200.16.10:40998.service - OpenSSH per-connection server daemon (10.200.16.10:40998). Mar 2 12:48:50.305479 sshd[2291]: Accepted publickey for core from 10.200.16.10 port 40998 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:48:50.306171 sshd-session[2291]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:50.309597 systemd-logind[1870]: New session 3 of user core. Mar 2 12:48:50.316617 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 2 12:48:50.620631 systemd[1]: Started sshd@1-10.200.20.38:22-10.200.16.10:52228.service - OpenSSH per-connection server daemon (10.200.16.10:52228). Mar 2 12:48:50.786593 chronyd[1850]: Selected source PHC0 Mar 2 12:48:51.034944 sshd[2297]: Accepted publickey for core from 10.200.16.10 port 52228 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:48:51.035987 sshd-session[2297]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:51.039390 systemd-logind[1870]: New session 4 of user core. Mar 2 12:48:51.047756 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 2 12:48:51.271216 sshd[2300]: Connection closed by 10.200.16.10 port 52228 Mar 2 12:48:51.271743 sshd-session[2297]: pam_unix(sshd:session): session closed for user core Mar 2 12:48:51.274877 systemd[1]: sshd@1-10.200.20.38:22-10.200.16.10:52228.service: Deactivated successfully. Mar 2 12:48:51.276485 systemd[1]: session-4.scope: Deactivated successfully. Mar 2 12:48:51.278622 systemd-logind[1870]: Session 4 logged out. Waiting for processes to exit. Mar 2 12:48:51.279481 systemd-logind[1870]: Removed session 4. Mar 2 12:48:51.357675 systemd[1]: Started sshd@2-10.200.20.38:22-10.200.16.10:52238.service - OpenSSH per-connection server daemon (10.200.16.10:52238). Mar 2 12:48:51.773878 sshd[2306]: Accepted publickey for core from 10.200.16.10 port 52238 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:48:51.774954 sshd-session[2306]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:51.778796 systemd-logind[1870]: New session 5 of user core. Mar 2 12:48:51.788600 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 2 12:48:52.004183 sshd[2309]: Connection closed by 10.200.16.10 port 52238 Mar 2 12:48:52.004588 sshd-session[2306]: pam_unix(sshd:session): session closed for user core Mar 2 12:48:52.007327 systemd-logind[1870]: Session 5 logged out. Waiting for processes to exit. Mar 2 12:48:52.007834 systemd[1]: sshd@2-10.200.20.38:22-10.200.16.10:52238.service: Deactivated successfully. Mar 2 12:48:52.009351 systemd[1]: session-5.scope: Deactivated successfully. Mar 2 12:48:52.010266 systemd-logind[1870]: Removed session 5. Mar 2 12:48:52.095525 systemd[1]: Started sshd@3-10.200.20.38:22-10.200.16.10:52240.service - OpenSSH per-connection server daemon (10.200.16.10:52240). Mar 2 12:48:52.520307 sshd[2315]: Accepted publickey for core from 10.200.16.10 port 52240 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:48:52.521370 sshd-session[2315]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:52.524503 systemd-logind[1870]: New session 6 of user core. Mar 2 12:48:52.531762 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 2 12:48:52.754913 sshd[2318]: Connection closed by 10.200.16.10 port 52240 Mar 2 12:48:52.755321 sshd-session[2315]: pam_unix(sshd:session): session closed for user core Mar 2 12:48:52.758405 systemd[1]: sshd@3-10.200.20.38:22-10.200.16.10:52240.service: Deactivated successfully. Mar 2 12:48:52.759922 systemd[1]: session-6.scope: Deactivated successfully. Mar 2 12:48:52.760550 systemd-logind[1870]: Session 6 logged out. Waiting for processes to exit. Mar 2 12:48:52.761371 systemd-logind[1870]: Removed session 6. Mar 2 12:48:52.843449 systemd[1]: Started sshd@4-10.200.20.38:22-10.200.16.10:52248.service - OpenSSH per-connection server daemon (10.200.16.10:52248). Mar 2 12:48:53.261278 sshd[2324]: Accepted publickey for core from 10.200.16.10 port 52248 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:48:53.262102 sshd-session[2324]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:53.265359 systemd-logind[1870]: New session 7 of user core. Mar 2 12:48:53.273815 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 2 12:48:53.780782 sudo[2328]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 2 12:48:53.780986 sudo[2328]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:48:53.793079 sudo[2328]: pam_unix(sudo:session): session closed for user root Mar 2 12:48:53.871083 sshd[2327]: Connection closed by 10.200.16.10 port 52248 Mar 2 12:48:53.870994 sshd-session[2324]: pam_unix(sshd:session): session closed for user core Mar 2 12:48:53.874849 systemd[1]: sshd@4-10.200.20.38:22-10.200.16.10:52248.service: Deactivated successfully. Mar 2 12:48:53.876684 systemd[1]: session-7.scope: Deactivated successfully. Mar 2 12:48:53.877640 systemd-logind[1870]: Session 7 logged out. Waiting for processes to exit. Mar 2 12:48:53.879102 systemd-logind[1870]: Removed session 7. Mar 2 12:48:53.958684 systemd[1]: Started sshd@5-10.200.20.38:22-10.200.16.10:52256.service - OpenSSH per-connection server daemon (10.200.16.10:52256). Mar 2 12:48:54.378648 sshd[2334]: Accepted publickey for core from 10.200.16.10 port 52256 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:48:54.379670 sshd-session[2334]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:54.383418 systemd-logind[1870]: New session 8 of user core. Mar 2 12:48:54.393593 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 2 12:48:54.535208 sudo[2339]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 2 12:48:54.535406 sudo[2339]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:48:54.542041 sudo[2339]: pam_unix(sudo:session): session closed for user root Mar 2 12:48:54.545297 sudo[2338]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 2 12:48:54.545476 sudo[2338]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:48:54.551566 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 2 12:48:54.577615 augenrules[2361]: No rules Mar 2 12:48:54.578631 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 12:48:54.578874 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 2 12:48:54.580764 sudo[2338]: pam_unix(sudo:session): session closed for user root Mar 2 12:48:54.658023 sshd[2337]: Connection closed by 10.200.16.10 port 52256 Mar 2 12:48:54.657806 sshd-session[2334]: pam_unix(sshd:session): session closed for user core Mar 2 12:48:54.660938 systemd[1]: sshd@5-10.200.20.38:22-10.200.16.10:52256.service: Deactivated successfully. Mar 2 12:48:54.662312 systemd[1]: session-8.scope: Deactivated successfully. Mar 2 12:48:54.663137 systemd-logind[1870]: Session 8 logged out. Waiting for processes to exit. Mar 2 12:48:54.664754 systemd-logind[1870]: Removed session 8. Mar 2 12:48:54.748798 systemd[1]: Started sshd@6-10.200.20.38:22-10.200.16.10:52260.service - OpenSSH per-connection server daemon (10.200.16.10:52260). Mar 2 12:48:55.168006 sshd[2370]: Accepted publickey for core from 10.200.16.10 port 52260 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:48:55.168910 sshd-session[2370]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:48:55.172254 systemd-logind[1870]: New session 9 of user core. Mar 2 12:48:55.180773 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 2 12:48:55.324332 sudo[2374]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 2 12:48:55.324795 sudo[2374]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:48:57.819271 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 2 12:48:57.830879 (dockerd)[2392]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 2 12:48:58.833004 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 2 12:48:58.834194 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:48:59.280357 dockerd[2392]: time="2026-03-02T12:48:59.280251270Z" level=info msg="Starting up" Mar 2 12:48:59.282092 dockerd[2392]: time="2026-03-02T12:48:59.282074007Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Mar 2 12:48:59.289742 dockerd[2392]: time="2026-03-02T12:48:59.289719608Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Mar 2 12:48:59.364147 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:48:59.370878 (kubelet)[2420]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:48:59.395843 kubelet[2420]: E0302 12:48:59.395796 2420 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:48:59.397575 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:48:59.397680 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:48:59.397931 systemd[1]: kubelet.service: Consumed 103ms CPU time, 106.9M memory peak. Mar 2 12:48:59.468090 dockerd[2392]: time="2026-03-02T12:48:59.468066798Z" level=info msg="Loading containers: start." Mar 2 12:48:59.524514 kernel: Initializing XFRM netlink socket Mar 2 12:48:59.918930 systemd-networkd[1485]: docker0: Link UP Mar 2 12:48:59.936513 dockerd[2392]: time="2026-03-02T12:48:59.936243687Z" level=info msg="Loading containers: done." Mar 2 12:48:59.952751 dockerd[2392]: time="2026-03-02T12:48:59.952721350Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 2 12:48:59.952859 dockerd[2392]: time="2026-03-02T12:48:59.952778106Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Mar 2 12:48:59.952859 dockerd[2392]: time="2026-03-02T12:48:59.952848498Z" level=info msg="Initializing buildkit" Mar 2 12:49:00.000314 dockerd[2392]: time="2026-03-02T12:49:00.000288751Z" level=info msg="Completed buildkit initialization" Mar 2 12:49:00.005309 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 2 12:49:00.005556 dockerd[2392]: time="2026-03-02T12:49:00.005459863Z" level=info msg="Daemon has completed initialization" Mar 2 12:49:00.006284 dockerd[2392]: time="2026-03-02T12:49:00.006244540Z" level=info msg="API listen on /run/docker.sock" Mar 2 12:49:00.284854 containerd[1899]: time="2026-03-02T12:49:00.284740626Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.2\"" Mar 2 12:49:01.223826 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3011676218.mount: Deactivated successfully. Mar 2 12:49:02.663524 containerd[1899]: time="2026-03-02T12:49:02.663296352Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:02.667283 containerd[1899]: time="2026-03-02T12:49:02.667253439Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.35.2: active requests=0, bytes read=24701796" Mar 2 12:49:02.670401 containerd[1899]: time="2026-03-02T12:49:02.670365267Z" level=info msg="ImageCreate event name:\"sha256:713a7d5fc5ed8383c9ffe550e487150c9818d05f0c4c012688fbb27885fcc7bf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:02.675140 containerd[1899]: time="2026-03-02T12:49:02.675118243Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:68cdc586f13b13edb7aa30a18155be530136a39cfd5ef8672aad8ccc98f0a7f7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:02.676270 containerd[1899]: time="2026-03-02T12:49:02.676246975Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.35.2\" with image id \"sha256:713a7d5fc5ed8383c9ffe550e487150c9818d05f0c4c012688fbb27885fcc7bf\", repo tag \"registry.k8s.io/kube-apiserver:v1.35.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:68cdc586f13b13edb7aa30a18155be530136a39cfd5ef8672aad8ccc98f0a7f7\", size \"24698395\" in 2.391469526s" Mar 2 12:49:02.676374 containerd[1899]: time="2026-03-02T12:49:02.676360626Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.2\" returns image reference \"sha256:713a7d5fc5ed8383c9ffe550e487150c9818d05f0c4c012688fbb27885fcc7bf\"" Mar 2 12:49:02.677013 containerd[1899]: time="2026-03-02T12:49:02.676850338Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.2\"" Mar 2 12:49:04.196480 containerd[1899]: time="2026-03-02T12:49:04.196382890Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:04.199523 containerd[1899]: time="2026-03-02T12:49:04.199500822Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.35.2: active requests=0, bytes read=19063039" Mar 2 12:49:04.202732 containerd[1899]: time="2026-03-02T12:49:04.202693932Z" level=info msg="ImageCreate event name:\"sha256:6137f51959af5f0a4da7fb6c0bd868f615a534c02d42e303ad6fb31345ee4854\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:04.207448 containerd[1899]: time="2026-03-02T12:49:04.207243286Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:d9784320a41dd1b155c0ad8fdb5823d60c475870f3dd23865edde36b585748f2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:04.207727 containerd[1899]: time="2026-03-02T12:49:04.207705565Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.35.2\" with image id \"sha256:6137f51959af5f0a4da7fb6c0bd868f615a534c02d42e303ad6fb31345ee4854\", repo tag \"registry.k8s.io/kube-controller-manager:v1.35.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:d9784320a41dd1b155c0ad8fdb5823d60c475870f3dd23865edde36b585748f2\", size \"20675140\" in 1.530612307s" Mar 2 12:49:04.207817 containerd[1899]: time="2026-03-02T12:49:04.207801480Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.2\" returns image reference \"sha256:6137f51959af5f0a4da7fb6c0bd868f615a534c02d42e303ad6fb31345ee4854\"" Mar 2 12:49:04.208317 containerd[1899]: time="2026-03-02T12:49:04.208277879Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.2\"" Mar 2 12:49:05.561527 containerd[1899]: time="2026-03-02T12:49:05.561185676Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:05.563962 containerd[1899]: time="2026-03-02T12:49:05.563938588Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.35.2: active requests=0, bytes read=13797901" Mar 2 12:49:05.567114 containerd[1899]: time="2026-03-02T12:49:05.567077008Z" level=info msg="ImageCreate event name:\"sha256:6ad431b09accba3ccc8ac6df4b239aa11c7adf8ee0a477b9f0b54cf9f083f8c6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:05.572081 containerd[1899]: time="2026-03-02T12:49:05.572048871Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:5833e2c4b779215efe7a48126c067de199e86aa5a86518693adeef16db0ff943\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:05.572696 containerd[1899]: time="2026-03-02T12:49:05.572670755Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.35.2\" with image id \"sha256:6ad431b09accba3ccc8ac6df4b239aa11c7adf8ee0a477b9f0b54cf9f083f8c6\", repo tag \"registry.k8s.io/kube-scheduler:v1.35.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:5833e2c4b779215efe7a48126c067de199e86aa5a86518693adeef16db0ff943\", size \"15410020\" in 1.364350779s" Mar 2 12:49:05.572722 containerd[1899]: time="2026-03-02T12:49:05.572699716Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.2\" returns image reference \"sha256:6ad431b09accba3ccc8ac6df4b239aa11c7adf8ee0a477b9f0b54cf9f083f8c6\"" Mar 2 12:49:05.573384 containerd[1899]: time="2026-03-02T12:49:05.573343201Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.2\"" Mar 2 12:49:06.768150 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2573762266.mount: Deactivated successfully. Mar 2 12:49:06.967185 containerd[1899]: time="2026-03-02T12:49:06.967130287Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.35.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:06.970892 containerd[1899]: time="2026-03-02T12:49:06.970865603Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.35.2: active requests=0, bytes read=22329583" Mar 2 12:49:06.973867 containerd[1899]: time="2026-03-02T12:49:06.973839507Z" level=info msg="ImageCreate event name:\"sha256:df7dcaf93e84e5dfbe96b2f86588b38a8959748d9c84b2e0532e2b5ae1bc5884\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:06.978136 containerd[1899]: time="2026-03-02T12:49:06.978109132Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:015265214cc874b593a7adccdcfe4ac15d2b8e9ae89881bdcd5bcb99d42e1862\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:06.978821 containerd[1899]: time="2026-03-02T12:49:06.978794622Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.35.2\" with image id \"sha256:df7dcaf93e84e5dfbe96b2f86588b38a8959748d9c84b2e0532e2b5ae1bc5884\", repo tag \"registry.k8s.io/kube-proxy:v1.35.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:015265214cc874b593a7adccdcfe4ac15d2b8e9ae89881bdcd5bcb99d42e1862\", size \"22328602\" in 1.405415851s" Mar 2 12:49:06.978843 containerd[1899]: time="2026-03-02T12:49:06.978826375Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.2\" returns image reference \"sha256:df7dcaf93e84e5dfbe96b2f86588b38a8959748d9c84b2e0532e2b5ae1bc5884\"" Mar 2 12:49:06.979390 containerd[1899]: time="2026-03-02T12:49:06.979370675Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\"" Mar 2 12:49:07.594304 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4059632523.mount: Deactivated successfully. Mar 2 12:49:08.872874 containerd[1899]: time="2026-03-02T12:49:08.872816499Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.13.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:08.877548 containerd[1899]: time="2026-03-02T12:49:08.877520428Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.13.1: active requests=0, bytes read=21172211" Mar 2 12:49:08.881755 containerd[1899]: time="2026-03-02T12:49:08.881715546Z" level=info msg="ImageCreate event name:\"sha256:e08f4d9d2e6ede8185064c13b41f8eeee95b609c0ca93b6fe7509fe527c907cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:08.886901 containerd[1899]: time="2026-03-02T12:49:08.886868748Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:08.887811 containerd[1899]: time="2026-03-02T12:49:08.887544341Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.13.1\" with image id \"sha256:e08f4d9d2e6ede8185064c13b41f8eeee95b609c0ca93b6fe7509fe527c907cf\", repo tag \"registry.k8s.io/coredns/coredns:v1.13.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\", size \"21168808\" in 1.908147809s" Mar 2 12:49:08.887811 containerd[1899]: time="2026-03-02T12:49:08.887571606Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\" returns image reference \"sha256:e08f4d9d2e6ede8185064c13b41f8eeee95b609c0ca93b6fe7509fe527c907cf\"" Mar 2 12:49:08.888109 containerd[1899]: time="2026-03-02T12:49:08.888053240Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Mar 2 12:49:09.445781 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 2 12:49:09.447325 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:49:09.452939 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2690112492.mount: Deactivated successfully. Mar 2 12:49:09.484836 containerd[1899]: time="2026-03-02T12:49:09.484791575Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:09.537425 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:49:09.541734 (kubelet)[2756]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:49:09.564575 kubelet[2756]: E0302 12:49:09.564537 2756 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:49:09.566263 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:49:09.566445 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:49:09.568552 systemd[1]: kubelet.service: Consumed 100ms CPU time, 107M memory peak. Mar 2 12:49:09.928018 containerd[1899]: time="2026-03-02T12:49:09.927791892Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=268709" Mar 2 12:49:09.992504 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Mar 2 12:49:09.993930 containerd[1899]: time="2026-03-02T12:49:09.993398831Z" level=info msg="ImageCreate event name:\"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:09.998944 containerd[1899]: time="2026-03-02T12:49:09.998919439Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:09.999287 containerd[1899]: time="2026-03-02T12:49:09.999260260Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"267939\" in 1.111080543s" Mar 2 12:49:09.999287 containerd[1899]: time="2026-03-02T12:49:09.999286661Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\"" Mar 2 12:49:10.000065 containerd[1899]: time="2026-03-02T12:49:10.000044354Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\"" Mar 2 12:49:11.005476 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2033208828.mount: Deactivated successfully. Mar 2 12:49:12.172944 containerd[1899]: time="2026-03-02T12:49:12.172893772Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.6-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:12.175859 containerd[1899]: time="2026-03-02T12:49:12.175835691Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.6-0: active requests=0, bytes read=21738165" Mar 2 12:49:12.179017 containerd[1899]: time="2026-03-02T12:49:12.178982377Z" level=info msg="ImageCreate event name:\"sha256:271e49a0ebc56647476845128fcd2a73bb138beeca3878cc3bf52b4ff1172a57\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:12.184516 containerd[1899]: time="2026-03-02T12:49:12.184308466Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:12.184794 containerd[1899]: time="2026-03-02T12:49:12.184771163Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.6-0\" with image id \"sha256:271e49a0ebc56647476845128fcd2a73bb138beeca3878cc3bf52b4ff1172a57\", repo tag \"registry.k8s.io/etcd:3.6.6-0\", repo digest \"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\", size \"21749640\" in 2.184623302s" Mar 2 12:49:12.184870 containerd[1899]: time="2026-03-02T12:49:12.184858262Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\" returns image reference \"sha256:271e49a0ebc56647476845128fcd2a73bb138beeca3878cc3bf52b4ff1172a57\"" Mar 2 12:49:12.487642 update_engine[1875]: I20260302 12:49:12.487517 1875 update_attempter.cc:509] Updating boot flags... Mar 2 12:49:13.266876 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:49:13.267167 systemd[1]: kubelet.service: Consumed 100ms CPU time, 107M memory peak. Mar 2 12:49:13.269697 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:49:13.292352 systemd[1]: Reload requested from client PID 2966 ('systemctl') (unit session-9.scope)... Mar 2 12:49:13.292367 systemd[1]: Reloading... Mar 2 12:49:13.378521 zram_generator::config[3010]: No configuration found. Mar 2 12:49:13.526893 systemd[1]: Reloading finished in 234 ms. Mar 2 12:49:13.576905 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 2 12:49:13.576966 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 2 12:49:13.578548 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:49:13.578588 systemd[1]: kubelet.service: Consumed 74ms CPU time, 94.9M memory peak. Mar 2 12:49:13.581691 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:49:13.812274 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:49:13.819684 (kubelet)[3080]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 12:49:13.843798 kubelet[3080]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 12:49:13.987748 kubelet[3080]: I0302 12:49:13.987611 3080 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Mar 2 12:49:13.987748 kubelet[3080]: I0302 12:49:13.987744 3080 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 12:49:13.988712 kubelet[3080]: I0302 12:49:13.988691 3080 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 2 12:49:13.988712 kubelet[3080]: I0302 12:49:13.988706 3080 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 12:49:13.988902 kubelet[3080]: I0302 12:49:13.988885 3080 server.go:951] "Client rotation is on, will bootstrap in background" Mar 2 12:49:14.350171 kubelet[3080]: E0302 12:49:14.350130 3080 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.200.20.38:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.20.38:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 12:49:14.350452 kubelet[3080]: I0302 12:49:14.350425 3080 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 12:49:14.353707 kubelet[3080]: I0302 12:49:14.353682 3080 server.go:1418] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 12:49:14.356395 kubelet[3080]: I0302 12:49:14.356379 3080 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 2 12:49:14.357086 kubelet[3080]: I0302 12:49:14.357055 3080 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 12:49:14.357218 kubelet[3080]: I0302 12:49:14.357086 3080 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459.2.101-47783670b7","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 12:49:14.357300 kubelet[3080]: I0302 12:49:14.357221 3080 topology_manager.go:143] "Creating topology manager with none policy" Mar 2 12:49:14.357300 kubelet[3080]: I0302 12:49:14.357228 3080 container_manager_linux.go:308] "Creating device plugin manager" Mar 2 12:49:14.357339 kubelet[3080]: I0302 12:49:14.357309 3080 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Mar 2 12:49:14.362516 kubelet[3080]: I0302 12:49:14.362479 3080 state_mem.go:41] "Initialized" logger="CPUManager state memory" Mar 2 12:49:14.362649 kubelet[3080]: I0302 12:49:14.362635 3080 kubelet.go:482] "Attempting to sync node with API server" Mar 2 12:49:14.362670 kubelet[3080]: I0302 12:49:14.362654 3080 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 12:49:14.363648 kubelet[3080]: I0302 12:49:14.362679 3080 kubelet.go:394] "Adding apiserver pod source" Mar 2 12:49:14.363648 kubelet[3080]: I0302 12:49:14.363540 3080 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 12:49:14.366553 kubelet[3080]: I0302 12:49:14.366539 3080 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 2 12:49:14.367288 kubelet[3080]: I0302 12:49:14.367273 3080 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 12:49:14.367399 kubelet[3080]: I0302 12:49:14.367390 3080 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 2 12:49:14.367472 kubelet[3080]: W0302 12:49:14.367464 3080 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 2 12:49:14.369534 kubelet[3080]: I0302 12:49:14.369389 3080 server.go:1257] "Started kubelet" Mar 2 12:49:14.373368 kubelet[3080]: I0302 12:49:14.373351 3080 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Mar 2 12:49:14.374812 kubelet[3080]: E0302 12:49:14.374115 3080 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.38:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.38:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4459.2.101-47783670b7.1899071e23486bf7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4459.2.101-47783670b7,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4459.2.101-47783670b7,},FirstTimestamp:2026-03-02 12:49:14.369362935 +0000 UTC m=+0.547105260,LastTimestamp:2026-03-02 12:49:14.369362935 +0000 UTC m=+0.547105260,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459.2.101-47783670b7,}" Mar 2 12:49:14.375968 kubelet[3080]: I0302 12:49:14.375933 3080 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 12:49:14.376836 kubelet[3080]: I0302 12:49:14.376822 3080 volume_manager.go:311] "Starting Kubelet Volume Manager" Mar 2 12:49:14.377070 kubelet[3080]: E0302 12:49:14.377052 3080 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459.2.101-47783670b7\" not found" Mar 2 12:49:14.378057 kubelet[3080]: I0302 12:49:14.378032 3080 server.go:317] "Adding debug handlers to kubelet server" Mar 2 12:49:14.380104 kubelet[3080]: I0302 12:49:14.380061 3080 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 12:49:14.380160 kubelet[3080]: I0302 12:49:14.380109 3080 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 2 12:49:14.380250 kubelet[3080]: I0302 12:49:14.380233 3080 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 12:49:14.381216 kubelet[3080]: E0302 12:49:14.380935 3080 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.101-47783670b7?timeout=10s\": dial tcp 10.200.20.38:6443: connect: connection refused" interval="200ms" Mar 2 12:49:14.381216 kubelet[3080]: I0302 12:49:14.376856 3080 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 12:49:14.381216 kubelet[3080]: I0302 12:49:14.381133 3080 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 2 12:49:14.381406 kubelet[3080]: I0302 12:49:14.381380 3080 reconciler.go:29] "Reconciler: start to sync state" Mar 2 12:49:14.382097 kubelet[3080]: I0302 12:49:14.382074 3080 factory.go:223] Registration of the systemd container factory successfully Mar 2 12:49:14.382152 kubelet[3080]: I0302 12:49:14.382137 3080 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 12:49:14.382443 kubelet[3080]: E0302 12:49:14.382421 3080 kubelet.go:1656] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 12:49:14.383258 kubelet[3080]: I0302 12:49:14.383240 3080 factory.go:223] Registration of the containerd container factory successfully Mar 2 12:49:14.401861 kubelet[3080]: I0302 12:49:14.401845 3080 cpu_manager.go:225] "Starting" policy="none" Mar 2 12:49:14.401861 kubelet[3080]: I0302 12:49:14.401855 3080 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 2 12:49:14.401946 kubelet[3080]: I0302 12:49:14.401874 3080 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Mar 2 12:49:14.408202 kubelet[3080]: I0302 12:49:14.408184 3080 policy_none.go:50] "Start" Mar 2 12:49:14.408202 kubelet[3080]: I0302 12:49:14.408201 3080 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 2 12:49:14.408285 kubelet[3080]: I0302 12:49:14.408210 3080 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 2 12:49:14.414592 kubelet[3080]: I0302 12:49:14.414571 3080 policy_none.go:44] "Start" Mar 2 12:49:14.418011 kubelet[3080]: I0302 12:49:14.417982 3080 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 2 12:49:14.419464 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 2 12:49:14.420674 kubelet[3080]: I0302 12:49:14.420651 3080 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 2 12:49:14.420674 kubelet[3080]: I0302 12:49:14.420669 3080 status_manager.go:249] "Starting to sync pod status with apiserver" Mar 2 12:49:14.420743 kubelet[3080]: I0302 12:49:14.420691 3080 kubelet.go:2501] "Starting kubelet main sync loop" Mar 2 12:49:14.420743 kubelet[3080]: E0302 12:49:14.420717 3080 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 12:49:14.432816 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 2 12:49:14.435410 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 2 12:49:14.448958 kubelet[3080]: E0302 12:49:14.448942 3080 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 12:49:14.449444 kubelet[3080]: I0302 12:49:14.449164 3080 eviction_manager.go:194] "Eviction manager: starting control loop" Mar 2 12:49:14.449444 kubelet[3080]: I0302 12:49:14.449180 3080 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 12:49:14.449444 kubelet[3080]: I0302 12:49:14.449389 3080 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Mar 2 12:49:14.451061 kubelet[3080]: E0302 12:49:14.450952 3080 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 12:49:14.451154 kubelet[3080]: E0302 12:49:14.451144 3080 eviction_manager.go:297] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4459.2.101-47783670b7\" not found" Mar 2 12:49:14.534179 systemd[1]: Created slice kubepods-burstable-pod3763a614b477a2af03685e55c26d0edd.slice - libcontainer container kubepods-burstable-pod3763a614b477a2af03685e55c26d0edd.slice. Mar 2 12:49:14.539280 kubelet[3080]: E0302 12:49:14.539257 3080 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-47783670b7\" not found" node="ci-4459.2.101-47783670b7" Mar 2 12:49:14.542381 systemd[1]: Created slice kubepods-burstable-pod6b349507c6e64e25503f34801fcd8580.slice - libcontainer container kubepods-burstable-pod6b349507c6e64e25503f34801fcd8580.slice. Mar 2 12:49:14.543894 kubelet[3080]: E0302 12:49:14.543875 3080 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-47783670b7\" not found" node="ci-4459.2.101-47783670b7" Mar 2 12:49:14.550929 kubelet[3080]: I0302 12:49:14.550905 3080 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459.2.101-47783670b7" Mar 2 12:49:14.551150 kubelet[3080]: E0302 12:49:14.551129 3080 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.200.20.38:6443/api/v1/nodes\": dial tcp 10.200.20.38:6443: connect: connection refused" node="ci-4459.2.101-47783670b7" Mar 2 12:49:14.553325 systemd[1]: Created slice kubepods-burstable-podf5e939b7da7aeee14418bdee8edf0b5e.slice - libcontainer container kubepods-burstable-podf5e939b7da7aeee14418bdee8edf0b5e.slice. Mar 2 12:49:14.554685 kubelet[3080]: E0302 12:49:14.554663 3080 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-47783670b7\" not found" node="ci-4459.2.101-47783670b7" Mar 2 12:49:14.582521 kubelet[3080]: I0302 12:49:14.582312 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3763a614b477a2af03685e55c26d0edd-ca-certs\") pod \"kube-apiserver-ci-4459.2.101-47783670b7\" (UID: \"3763a614b477a2af03685e55c26d0edd\") " pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582521 kubelet[3080]: I0302 12:49:14.582336 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3763a614b477a2af03685e55c26d0edd-k8s-certs\") pod \"kube-apiserver-ci-4459.2.101-47783670b7\" (UID: \"3763a614b477a2af03685e55c26d0edd\") " pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582521 kubelet[3080]: I0302 12:49:14.582354 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-ca-certs\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582521 kubelet[3080]: I0302 12:49:14.582367 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-kubeconfig\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582521 kubelet[3080]: I0302 12:49:14.582380 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582655 kubelet[3080]: I0302 12:49:14.582392 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3763a614b477a2af03685e55c26d0edd-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459.2.101-47783670b7\" (UID: \"3763a614b477a2af03685e55c26d0edd\") " pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582655 kubelet[3080]: I0302 12:49:14.582404 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-flexvolume-dir\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582655 kubelet[3080]: I0302 12:49:14.582418 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-k8s-certs\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582655 kubelet[3080]: I0302 12:49:14.582431 3080 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f5e939b7da7aeee14418bdee8edf0b5e-kubeconfig\") pod \"kube-scheduler-ci-4459.2.101-47783670b7\" (UID: \"f5e939b7da7aeee14418bdee8edf0b5e\") " pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" Mar 2 12:49:14.582712 kubelet[3080]: E0302 12:49:14.582660 3080 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.101-47783670b7?timeout=10s\": dial tcp 10.200.20.38:6443: connect: connection refused" interval="400ms" Mar 2 12:49:14.752660 kubelet[3080]: I0302 12:49:14.752588 3080 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459.2.101-47783670b7" Mar 2 12:49:14.752873 kubelet[3080]: E0302 12:49:14.752842 3080 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.200.20.38:6443/api/v1/nodes\": dial tcp 10.200.20.38:6443: connect: connection refused" node="ci-4459.2.101-47783670b7" Mar 2 12:49:14.849776 containerd[1899]: time="2026-03-02T12:49:14.849736900Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459.2.101-47783670b7,Uid:3763a614b477a2af03685e55c26d0edd,Namespace:kube-system,Attempt:0,}" Mar 2 12:49:14.854697 containerd[1899]: time="2026-03-02T12:49:14.854669922Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459.2.101-47783670b7,Uid:6b349507c6e64e25503f34801fcd8580,Namespace:kube-system,Attempt:0,}" Mar 2 12:49:14.860460 containerd[1899]: time="2026-03-02T12:49:14.860437698Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459.2.101-47783670b7,Uid:f5e939b7da7aeee14418bdee8edf0b5e,Namespace:kube-system,Attempt:0,}" Mar 2 12:49:14.983760 kubelet[3080]: E0302 12:49:14.983726 3080 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.101-47783670b7?timeout=10s\": dial tcp 10.200.20.38:6443: connect: connection refused" interval="800ms" Mar 2 12:49:15.154727 kubelet[3080]: I0302 12:49:15.154697 3080 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459.2.101-47783670b7" Mar 2 12:49:15.155020 kubelet[3080]: E0302 12:49:15.154998 3080 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://10.200.20.38:6443/api/v1/nodes\": dial tcp 10.200.20.38:6443: connect: connection refused" node="ci-4459.2.101-47783670b7" Mar 2 12:49:15.462257 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2073749640.mount: Deactivated successfully. Mar 2 12:49:15.488298 containerd[1899]: time="2026-03-02T12:49:15.488260485Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:49:15.501014 containerd[1899]: time="2026-03-02T12:49:15.500862519Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Mar 2 12:49:15.504919 containerd[1899]: time="2026-03-02T12:49:15.504892376Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:49:15.511526 containerd[1899]: time="2026-03-02T12:49:15.511152804Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:49:15.517763 containerd[1899]: time="2026-03-02T12:49:15.517122860Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 2 12:49:15.520572 containerd[1899]: time="2026-03-02T12:49:15.520545109Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:49:15.524102 containerd[1899]: time="2026-03-02T12:49:15.524078115Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:49:15.524644 containerd[1899]: time="2026-03-02T12:49:15.524623081Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 670.400288ms" Mar 2 12:49:15.527003 containerd[1899]: time="2026-03-02T12:49:15.526980367Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 2 12:49:15.530526 containerd[1899]: time="2026-03-02T12:49:15.530479060Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 668.798224ms" Mar 2 12:49:15.548177 containerd[1899]: time="2026-03-02T12:49:15.548145297Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 683.161129ms" Mar 2 12:49:15.565003 containerd[1899]: time="2026-03-02T12:49:15.564939299Z" level=info msg="connecting to shim f369bde20fb3075cdcebfdf3239cf83ecd9bf8b0d6be9644f54096a75e9346b2" address="unix:///run/containerd/s/33d2f17f2912384ed2bb18a34ae449d19ce138120d5d1b1e18b75809a89be684" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:49:15.584630 systemd[1]: Started cri-containerd-f369bde20fb3075cdcebfdf3239cf83ecd9bf8b0d6be9644f54096a75e9346b2.scope - libcontainer container f369bde20fb3075cdcebfdf3239cf83ecd9bf8b0d6be9644f54096a75e9346b2. Mar 2 12:49:15.612953 containerd[1899]: time="2026-03-02T12:49:15.612896256Z" level=info msg="connecting to shim c3404d425d00e3537006346dd6459634fb7d90cd5779e494a6b4317c9c5adb66" address="unix:///run/containerd/s/19fa30157a5343cbf632f0f568d1b6c9f2b94b501c27ef22b1c89dbedd52e3cb" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:49:15.624936 containerd[1899]: time="2026-03-02T12:49:15.624887522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459.2.101-47783670b7,Uid:3763a614b477a2af03685e55c26d0edd,Namespace:kube-system,Attempt:0,} returns sandbox id \"f369bde20fb3075cdcebfdf3239cf83ecd9bf8b0d6be9644f54096a75e9346b2\"" Mar 2 12:49:15.630744 containerd[1899]: time="2026-03-02T12:49:15.630341965Z" level=info msg="connecting to shim c4e1e0abd2f2d7168feaf4100192a84d963890c8f111f431ae29b84b153a6712" address="unix:///run/containerd/s/4d205fe41b6bc687ccfe08ea780d75d22839797852d367a6fe430e8e0217350f" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:49:15.632698 systemd[1]: Started cri-containerd-c3404d425d00e3537006346dd6459634fb7d90cd5779e494a6b4317c9c5adb66.scope - libcontainer container c3404d425d00e3537006346dd6459634fb7d90cd5779e494a6b4317c9c5adb66. Mar 2 12:49:15.635215 containerd[1899]: time="2026-03-02T12:49:15.635188559Z" level=info msg="CreateContainer within sandbox \"f369bde20fb3075cdcebfdf3239cf83ecd9bf8b0d6be9644f54096a75e9346b2\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 2 12:49:15.650610 systemd[1]: Started cri-containerd-c4e1e0abd2f2d7168feaf4100192a84d963890c8f111f431ae29b84b153a6712.scope - libcontainer container c4e1e0abd2f2d7168feaf4100192a84d963890c8f111f431ae29b84b153a6712. Mar 2 12:49:15.660327 containerd[1899]: time="2026-03-02T12:49:15.660288215Z" level=info msg="Container dca746428b2d5820cdd71be58dd3ee3985c14d7dc1be7bf61b823fbc992d4021: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:15.681499 containerd[1899]: time="2026-03-02T12:49:15.680806247Z" level=info msg="CreateContainer within sandbox \"f369bde20fb3075cdcebfdf3239cf83ecd9bf8b0d6be9644f54096a75e9346b2\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"dca746428b2d5820cdd71be58dd3ee3985c14d7dc1be7bf61b823fbc992d4021\"" Mar 2 12:49:15.682599 containerd[1899]: time="2026-03-02T12:49:15.682568181Z" level=info msg="StartContainer for \"dca746428b2d5820cdd71be58dd3ee3985c14d7dc1be7bf61b823fbc992d4021\"" Mar 2 12:49:15.683377 containerd[1899]: time="2026-03-02T12:49:15.683348765Z" level=info msg="connecting to shim dca746428b2d5820cdd71be58dd3ee3985c14d7dc1be7bf61b823fbc992d4021" address="unix:///run/containerd/s/33d2f17f2912384ed2bb18a34ae449d19ce138120d5d1b1e18b75809a89be684" protocol=ttrpc version=3 Mar 2 12:49:15.685147 containerd[1899]: time="2026-03-02T12:49:15.685065754Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459.2.101-47783670b7,Uid:6b349507c6e64e25503f34801fcd8580,Namespace:kube-system,Attempt:0,} returns sandbox id \"c3404d425d00e3537006346dd6459634fb7d90cd5779e494a6b4317c9c5adb66\"" Mar 2 12:49:15.693849 containerd[1899]: time="2026-03-02T12:49:15.693826425Z" level=info msg="CreateContainer within sandbox \"c3404d425d00e3537006346dd6459634fb7d90cd5779e494a6b4317c9c5adb66\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 2 12:49:15.702611 systemd[1]: Started cri-containerd-dca746428b2d5820cdd71be58dd3ee3985c14d7dc1be7bf61b823fbc992d4021.scope - libcontainer container dca746428b2d5820cdd71be58dd3ee3985c14d7dc1be7bf61b823fbc992d4021. Mar 2 12:49:15.705742 containerd[1899]: time="2026-03-02T12:49:15.705709710Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459.2.101-47783670b7,Uid:f5e939b7da7aeee14418bdee8edf0b5e,Namespace:kube-system,Attempt:0,} returns sandbox id \"c4e1e0abd2f2d7168feaf4100192a84d963890c8f111f431ae29b84b153a6712\"" Mar 2 12:49:15.715739 containerd[1899]: time="2026-03-02T12:49:15.715670590Z" level=info msg="CreateContainer within sandbox \"c4e1e0abd2f2d7168feaf4100192a84d963890c8f111f431ae29b84b153a6712\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 2 12:49:15.719105 containerd[1899]: time="2026-03-02T12:49:15.718641238Z" level=info msg="Container 6ac54abe142874ce2c496e6b102a0e2586aa44672151dfafcf73c59b5d2ceb4e: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:15.742600 containerd[1899]: time="2026-03-02T12:49:15.742575230Z" level=info msg="StartContainer for \"dca746428b2d5820cdd71be58dd3ee3985c14d7dc1be7bf61b823fbc992d4021\" returns successfully" Mar 2 12:49:15.744903 containerd[1899]: time="2026-03-02T12:49:15.744879163Z" level=info msg="CreateContainer within sandbox \"c3404d425d00e3537006346dd6459634fb7d90cd5779e494a6b4317c9c5adb66\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"6ac54abe142874ce2c496e6b102a0e2586aa44672151dfafcf73c59b5d2ceb4e\"" Mar 2 12:49:15.745552 containerd[1899]: time="2026-03-02T12:49:15.745535149Z" level=info msg="StartContainer for \"6ac54abe142874ce2c496e6b102a0e2586aa44672151dfafcf73c59b5d2ceb4e\"" Mar 2 12:49:15.747485 containerd[1899]: time="2026-03-02T12:49:15.747387904Z" level=info msg="connecting to shim 6ac54abe142874ce2c496e6b102a0e2586aa44672151dfafcf73c59b5d2ceb4e" address="unix:///run/containerd/s/19fa30157a5343cbf632f0f568d1b6c9f2b94b501c27ef22b1c89dbedd52e3cb" protocol=ttrpc version=3 Mar 2 12:49:15.755199 containerd[1899]: time="2026-03-02T12:49:15.755151047Z" level=info msg="Container ed2f1d2eba077ea550e83871461fc0d97d6d6930eeca0a02c819429ba7417bf0: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:15.770609 systemd[1]: Started cri-containerd-6ac54abe142874ce2c496e6b102a0e2586aa44672151dfafcf73c59b5d2ceb4e.scope - libcontainer container 6ac54abe142874ce2c496e6b102a0e2586aa44672151dfafcf73c59b5d2ceb4e. Mar 2 12:49:15.778787 containerd[1899]: time="2026-03-02T12:49:15.778713257Z" level=info msg="CreateContainer within sandbox \"c4e1e0abd2f2d7168feaf4100192a84d963890c8f111f431ae29b84b153a6712\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ed2f1d2eba077ea550e83871461fc0d97d6d6930eeca0a02c819429ba7417bf0\"" Mar 2 12:49:15.779123 containerd[1899]: time="2026-03-02T12:49:15.779101873Z" level=info msg="StartContainer for \"ed2f1d2eba077ea550e83871461fc0d97d6d6930eeca0a02c819429ba7417bf0\"" Mar 2 12:49:15.780563 containerd[1899]: time="2026-03-02T12:49:15.780511433Z" level=info msg="connecting to shim ed2f1d2eba077ea550e83871461fc0d97d6d6930eeca0a02c819429ba7417bf0" address="unix:///run/containerd/s/4d205fe41b6bc687ccfe08ea780d75d22839797852d367a6fe430e8e0217350f" protocol=ttrpc version=3 Mar 2 12:49:15.799613 systemd[1]: Started cri-containerd-ed2f1d2eba077ea550e83871461fc0d97d6d6930eeca0a02c819429ba7417bf0.scope - libcontainer container ed2f1d2eba077ea550e83871461fc0d97d6d6930eeca0a02c819429ba7417bf0. Mar 2 12:49:15.813813 containerd[1899]: time="2026-03-02T12:49:15.813772153Z" level=info msg="StartContainer for \"6ac54abe142874ce2c496e6b102a0e2586aa44672151dfafcf73c59b5d2ceb4e\" returns successfully" Mar 2 12:49:15.882315 containerd[1899]: time="2026-03-02T12:49:15.882229461Z" level=info msg="StartContainer for \"ed2f1d2eba077ea550e83871461fc0d97d6d6930eeca0a02c819429ba7417bf0\" returns successfully" Mar 2 12:49:15.956625 kubelet[3080]: I0302 12:49:15.956602 3080 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459.2.101-47783670b7" Mar 2 12:49:16.432417 kubelet[3080]: E0302 12:49:16.432135 3080 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-47783670b7\" not found" node="ci-4459.2.101-47783670b7" Mar 2 12:49:16.434264 kubelet[3080]: E0302 12:49:16.434232 3080 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-47783670b7\" not found" node="ci-4459.2.101-47783670b7" Mar 2 12:49:16.436790 kubelet[3080]: E0302 12:49:16.436638 3080 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-47783670b7\" not found" node="ci-4459.2.101-47783670b7" Mar 2 12:49:16.577218 kubelet[3080]: E0302 12:49:16.577180 3080 nodelease.go:50] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4459.2.101-47783670b7\" not found" node="ci-4459.2.101-47783670b7" Mar 2 12:49:17.038070 kubelet[3080]: I0302 12:49:17.037346 3080 kubelet_node_status.go:77] "Successfully registered node" node="ci-4459.2.101-47783670b7" Mar 2 12:49:17.077819 kubelet[3080]: I0302 12:49:17.077597 3080 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:17.136908 kubelet[3080]: E0302 12:49:17.136884 3080 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.101-47783670b7\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:17.137719 kubelet[3080]: I0302 12:49:17.137121 3080 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:17.139750 kubelet[3080]: E0302 12:49:17.139727 3080 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459.2.101-47783670b7\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:17.139750 kubelet[3080]: I0302 12:49:17.139746 3080 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" Mar 2 12:49:17.141537 kubelet[3080]: E0302 12:49:17.141514 3080 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459.2.101-47783670b7\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" Mar 2 12:49:17.367024 kubelet[3080]: I0302 12:49:17.367003 3080 apiserver.go:52] "Watching apiserver" Mar 2 12:49:17.381840 kubelet[3080]: I0302 12:49:17.381816 3080 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 2 12:49:17.437955 kubelet[3080]: I0302 12:49:17.437878 3080 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" Mar 2 12:49:17.438638 kubelet[3080]: I0302 12:49:17.438456 3080 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:17.440450 kubelet[3080]: E0302 12:49:17.440382 3080 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.101-47783670b7\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:17.441475 kubelet[3080]: E0302 12:49:17.441380 3080 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459.2.101-47783670b7\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" Mar 2 12:49:18.868649 systemd[1]: Reload requested from client PID 3364 ('systemctl') (unit session-9.scope)... Mar 2 12:49:18.868661 systemd[1]: Reloading... Mar 2 12:49:18.933538 zram_generator::config[3411]: No configuration found. Mar 2 12:49:19.091123 systemd[1]: Reloading finished in 222 ms. Mar 2 12:49:19.111329 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:49:19.125084 systemd[1]: kubelet.service: Deactivated successfully. Mar 2 12:49:19.125275 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:49:19.125321 systemd[1]: kubelet.service: Consumed 415ms CPU time, 122.3M memory peak. Mar 2 12:49:19.126566 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:49:19.286556 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:49:19.294727 (kubelet)[3475]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 12:49:19.322186 kubelet[3475]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 12:49:19.326595 kubelet[3475]: I0302 12:49:19.326556 3475 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Mar 2 12:49:19.326595 kubelet[3475]: I0302 12:49:19.326590 3475 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 12:49:19.326674 kubelet[3475]: I0302 12:49:19.326608 3475 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 2 12:49:19.326674 kubelet[3475]: I0302 12:49:19.326613 3475 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 12:49:19.326794 kubelet[3475]: I0302 12:49:19.326778 3475 server.go:951] "Client rotation is on, will bootstrap in background" Mar 2 12:49:19.327602 kubelet[3475]: I0302 12:49:19.327584 3475 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 2 12:49:19.329659 kubelet[3475]: I0302 12:49:19.329545 3475 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 12:49:19.334040 kubelet[3475]: I0302 12:49:19.334017 3475 server.go:1418] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 12:49:19.336365 kubelet[3475]: I0302 12:49:19.336346 3475 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 2 12:49:19.336539 kubelet[3475]: I0302 12:49:19.336506 3475 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 12:49:19.336639 kubelet[3475]: I0302 12:49:19.336536 3475 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459.2.101-47783670b7","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 12:49:19.336702 kubelet[3475]: I0302 12:49:19.336640 3475 topology_manager.go:143] "Creating topology manager with none policy" Mar 2 12:49:19.336702 kubelet[3475]: I0302 12:49:19.336645 3475 container_manager_linux.go:308] "Creating device plugin manager" Mar 2 12:49:19.336702 kubelet[3475]: I0302 12:49:19.336665 3475 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Mar 2 12:49:19.336853 kubelet[3475]: I0302 12:49:19.336836 3475 state_mem.go:41] "Initialized" logger="CPUManager state memory" Mar 2 12:49:19.336986 kubelet[3475]: I0302 12:49:19.336973 3475 kubelet.go:482] "Attempting to sync node with API server" Mar 2 12:49:19.337015 kubelet[3475]: I0302 12:49:19.336989 3475 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 12:49:19.337015 kubelet[3475]: I0302 12:49:19.337002 3475 kubelet.go:394] "Adding apiserver pod source" Mar 2 12:49:19.337051 kubelet[3475]: I0302 12:49:19.337018 3475 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 12:49:19.339826 kubelet[3475]: I0302 12:49:19.338750 3475 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 2 12:49:19.339826 kubelet[3475]: I0302 12:49:19.339295 3475 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 12:49:19.339826 kubelet[3475]: I0302 12:49:19.339315 3475 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 2 12:49:19.344584 kubelet[3475]: I0302 12:49:19.344566 3475 server.go:1257] "Started kubelet" Mar 2 12:49:19.347095 kubelet[3475]: I0302 12:49:19.347067 3475 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Mar 2 12:49:19.362043 kubelet[3475]: I0302 12:49:19.361998 3475 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 12:49:19.362386 kubelet[3475]: I0302 12:49:19.362338 3475 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 12:49:19.362562 kubelet[3475]: I0302 12:49:19.362549 3475 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 2 12:49:19.363159 kubelet[3475]: I0302 12:49:19.363125 3475 server.go:317] "Adding debug handlers to kubelet server" Mar 2 12:49:19.366286 kubelet[3475]: I0302 12:49:19.366266 3475 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 12:49:19.367532 kubelet[3475]: I0302 12:49:19.367510 3475 volume_manager.go:311] "Starting Kubelet Volume Manager" Mar 2 12:49:19.368719 kubelet[3475]: I0302 12:49:19.368704 3475 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 12:49:19.369325 kubelet[3475]: I0302 12:49:19.369178 3475 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 2 12:49:19.369325 kubelet[3475]: I0302 12:49:19.369294 3475 reconciler.go:29] "Reconciler: start to sync state" Mar 2 12:49:19.370220 kubelet[3475]: I0302 12:49:19.370200 3475 factory.go:223] Registration of the systemd container factory successfully Mar 2 12:49:19.371020 kubelet[3475]: I0302 12:49:19.370998 3475 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 12:49:19.372281 kubelet[3475]: I0302 12:49:19.371939 3475 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 2 12:49:19.372415 kubelet[3475]: E0302 12:49:19.372400 3475 kubelet.go:1656] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 12:49:19.373080 kubelet[3475]: I0302 12:49:19.373065 3475 factory.go:223] Registration of the containerd container factory successfully Mar 2 12:49:19.379766 kubelet[3475]: I0302 12:49:19.379109 3475 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 2 12:49:19.379871 kubelet[3475]: I0302 12:49:19.379857 3475 status_manager.go:249] "Starting to sync pod status with apiserver" Mar 2 12:49:19.379931 kubelet[3475]: I0302 12:49:19.379924 3475 kubelet.go:2501] "Starting kubelet main sync loop" Mar 2 12:49:19.380009 kubelet[3475]: E0302 12:49:19.379990 3475 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 12:49:19.409999 kubelet[3475]: I0302 12:49:19.409986 3475 cpu_manager.go:225] "Starting" policy="none" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410264 3475 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410286 3475 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410367 3475 state_mem.go:94] "Updated default CPUSet" logger="CPUManager state checkpoint.CPUManager state memory" cpuSet="" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410375 3475 state_mem.go:102] "Updated CPUSet assignments" logger="CPUManager state checkpoint.CPUManager state memory" assignments={} Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410385 3475 policy_none.go:50] "Start" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410391 3475 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410397 3475 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410457 3475 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Mar 2 12:49:19.410534 kubelet[3475]: I0302 12:49:19.410467 3475 policy_none.go:44] "Start" Mar 2 12:49:19.413437 kubelet[3475]: E0302 12:49:19.413423 3475 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 12:49:19.413655 kubelet[3475]: I0302 12:49:19.413641 3475 eviction_manager.go:194] "Eviction manager: starting control loop" Mar 2 12:49:19.414009 kubelet[3475]: I0302 12:49:19.413923 3475 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 12:49:19.415962 kubelet[3475]: I0302 12:49:19.415942 3475 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Mar 2 12:49:19.418174 kubelet[3475]: E0302 12:49:19.417253 3475 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 12:49:19.481392 kubelet[3475]: I0302 12:49:19.481369 3475 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" Mar 2 12:49:19.481789 kubelet[3475]: I0302 12:49:19.481604 3475 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:19.481875 kubelet[3475]: I0302 12:49:19.481670 3475 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:19.488972 kubelet[3475]: I0302 12:49:19.488951 3475 warnings.go:107] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:49:19.494224 kubelet[3475]: I0302 12:49:19.494202 3475 warnings.go:107] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:49:19.494653 kubelet[3475]: I0302 12:49:19.494356 3475 warnings.go:107] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:49:19.520457 kubelet[3475]: I0302 12:49:19.520440 3475 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459.2.101-47783670b7" Mar 2 12:49:19.539317 kubelet[3475]: I0302 12:49:19.539283 3475 kubelet_node_status.go:123] "Node was previously registered" node="ci-4459.2.101-47783670b7" Mar 2 12:49:19.539392 kubelet[3475]: I0302 12:49:19.539344 3475 kubelet_node_status.go:77] "Successfully registered node" node="ci-4459.2.101-47783670b7" Mar 2 12:49:19.571509 kubelet[3475]: I0302 12:49:19.570055 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3763a614b477a2af03685e55c26d0edd-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459.2.101-47783670b7\" (UID: \"3763a614b477a2af03685e55c26d0edd\") " pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571509 kubelet[3475]: I0302 12:49:19.570086 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-flexvolume-dir\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571509 kubelet[3475]: I0302 12:49:19.570099 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-k8s-certs\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571509 kubelet[3475]: I0302 12:49:19.570117 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3763a614b477a2af03685e55c26d0edd-ca-certs\") pod \"kube-apiserver-ci-4459.2.101-47783670b7\" (UID: \"3763a614b477a2af03685e55c26d0edd\") " pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571509 kubelet[3475]: I0302 12:49:19.570168 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3763a614b477a2af03685e55c26d0edd-k8s-certs\") pod \"kube-apiserver-ci-4459.2.101-47783670b7\" (UID: \"3763a614b477a2af03685e55c26d0edd\") " pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571680 kubelet[3475]: I0302 12:49:19.570197 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-ca-certs\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571680 kubelet[3475]: I0302 12:49:19.570212 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-kubeconfig\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571680 kubelet[3475]: I0302 12:49:19.570311 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6b349507c6e64e25503f34801fcd8580-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459.2.101-47783670b7\" (UID: \"6b349507c6e64e25503f34801fcd8580\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" Mar 2 12:49:19.571680 kubelet[3475]: I0302 12:49:19.570325 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f5e939b7da7aeee14418bdee8edf0b5e-kubeconfig\") pod \"kube-scheduler-ci-4459.2.101-47783670b7\" (UID: \"f5e939b7da7aeee14418bdee8edf0b5e\") " pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" Mar 2 12:49:20.337894 kubelet[3475]: I0302 12:49:20.337708 3475 apiserver.go:52] "Watching apiserver" Mar 2 12:49:20.369473 kubelet[3475]: I0302 12:49:20.369420 3475 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 2 12:49:20.402371 kubelet[3475]: I0302 12:49:20.402338 3475 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:20.417031 kubelet[3475]: I0302 12:49:20.416300 3475 warnings.go:107] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:49:20.417181 kubelet[3475]: E0302 12:49:20.417164 3475 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.101-47783670b7\" already exists" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" Mar 2 12:49:20.435442 kubelet[3475]: I0302 12:49:20.435323 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4459.2.101-47783670b7" podStartSLOduration=1.435306852 podStartE2EDuration="1.435306852s" podCreationTimestamp="2026-03-02 12:49:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:49:20.422232632 +0000 UTC m=+1.124563371" watchObservedRunningTime="2026-03-02 12:49:20.435306852 +0000 UTC m=+1.137637599" Mar 2 12:49:20.435442 kubelet[3475]: I0302 12:49:20.435382 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4459.2.101-47783670b7" podStartSLOduration=1.435379455 podStartE2EDuration="1.435379455s" podCreationTimestamp="2026-03-02 12:49:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:49:20.435196976 +0000 UTC m=+1.137527715" watchObservedRunningTime="2026-03-02 12:49:20.435379455 +0000 UTC m=+1.137710194" Mar 2 12:49:20.448976 kubelet[3475]: I0302 12:49:20.448890 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4459.2.101-47783670b7" podStartSLOduration=1.448885189 podStartE2EDuration="1.448885189s" podCreationTimestamp="2026-03-02 12:49:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:49:20.448821395 +0000 UTC m=+1.151152134" watchObservedRunningTime="2026-03-02 12:49:20.448885189 +0000 UTC m=+1.151215928" Mar 2 12:49:25.674700 kubelet[3475]: I0302 12:49:25.674662 3475 kuberuntime_manager.go:2062] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 2 12:49:25.675466 containerd[1899]: time="2026-03-02T12:49:25.675416622Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 2 12:49:25.675943 kubelet[3475]: I0302 12:49:25.675600 3475 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 2 12:49:26.813227 systemd[1]: Created slice kubepods-besteffort-pod8de2c13d_9c71_4c2f_9b70_9fca4b5e8fd2.slice - libcontainer container kubepods-besteffort-pod8de2c13d_9c71_4c2f_9b70_9fca4b5e8fd2.slice. Mar 2 12:49:26.821661 kubelet[3475]: I0302 12:49:26.821568 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2-lib-modules\") pod \"kube-proxy-kk9tg\" (UID: \"8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2\") " pod="kube-system/kube-proxy-kk9tg" Mar 2 12:49:26.821661 kubelet[3475]: I0302 12:49:26.821600 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9df25\" (UniqueName: \"kubernetes.io/projected/8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2-kube-api-access-9df25\") pod \"kube-proxy-kk9tg\" (UID: \"8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2\") " pod="kube-system/kube-proxy-kk9tg" Mar 2 12:49:26.821661 kubelet[3475]: I0302 12:49:26.821612 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2-kube-proxy\") pod \"kube-proxy-kk9tg\" (UID: \"8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2\") " pod="kube-system/kube-proxy-kk9tg" Mar 2 12:49:26.821661 kubelet[3475]: I0302 12:49:26.821621 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2-xtables-lock\") pod \"kube-proxy-kk9tg\" (UID: \"8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2\") " pod="kube-system/kube-proxy-kk9tg" Mar 2 12:49:26.942691 systemd[1]: Created slice kubepods-besteffort-pod58fec222_c058_4252_86f6_3f6f395545c1.slice - libcontainer container kubepods-besteffort-pod58fec222_c058_4252_86f6_3f6f395545c1.slice. Mar 2 12:49:27.023223 kubelet[3475]: I0302 12:49:27.023159 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcsvs\" (UniqueName: \"kubernetes.io/projected/58fec222-c058-4252-86f6-3f6f395545c1-kube-api-access-tcsvs\") pod \"tigera-operator-6447996989-jxgk4\" (UID: \"58fec222-c058-4252-86f6-3f6f395545c1\") " pod="tigera-operator/tigera-operator-6447996989-jxgk4" Mar 2 12:49:27.023223 kubelet[3475]: I0302 12:49:27.023189 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/58fec222-c058-4252-86f6-3f6f395545c1-var-lib-calico\") pod \"tigera-operator-6447996989-jxgk4\" (UID: \"58fec222-c058-4252-86f6-3f6f395545c1\") " pod="tigera-operator/tigera-operator-6447996989-jxgk4" Mar 2 12:49:27.132584 containerd[1899]: time="2026-03-02T12:49:27.132533118Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kk9tg,Uid:8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2,Namespace:kube-system,Attempt:0,}" Mar 2 12:49:27.166883 containerd[1899]: time="2026-03-02T12:49:27.166858007Z" level=info msg="connecting to shim 6ade71f71e0352295a2dabdaec269237d39a21a0df58b63cf0596766256b9e97" address="unix:///run/containerd/s/4a7e3d152ae6384c296564c95881547677bacdc40e34c8766eafd26188fab70d" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:49:27.191604 systemd[1]: Started cri-containerd-6ade71f71e0352295a2dabdaec269237d39a21a0df58b63cf0596766256b9e97.scope - libcontainer container 6ade71f71e0352295a2dabdaec269237d39a21a0df58b63cf0596766256b9e97. Mar 2 12:49:27.216372 containerd[1899]: time="2026-03-02T12:49:27.216339016Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kk9tg,Uid:8de2c13d-9c71-4c2f-9b70-9fca4b5e8fd2,Namespace:kube-system,Attempt:0,} returns sandbox id \"6ade71f71e0352295a2dabdaec269237d39a21a0df58b63cf0596766256b9e97\"" Mar 2 12:49:27.225341 containerd[1899]: time="2026-03-02T12:49:27.225312820Z" level=info msg="CreateContainer within sandbox \"6ade71f71e0352295a2dabdaec269237d39a21a0df58b63cf0596766256b9e97\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 2 12:49:27.246965 containerd[1899]: time="2026-03-02T12:49:27.246939018Z" level=info msg="Container fa285ca7266eb82cd1811a797c32d6ce21ff62b46af51bf68aea6f9577068720: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:27.252255 containerd[1899]: time="2026-03-02T12:49:27.252222023Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6447996989-jxgk4,Uid:58fec222-c058-4252-86f6-3f6f395545c1,Namespace:tigera-operator,Attempt:0,}" Mar 2 12:49:27.265946 containerd[1899]: time="2026-03-02T12:49:27.265915197Z" level=info msg="CreateContainer within sandbox \"6ade71f71e0352295a2dabdaec269237d39a21a0df58b63cf0596766256b9e97\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"fa285ca7266eb82cd1811a797c32d6ce21ff62b46af51bf68aea6f9577068720\"" Mar 2 12:49:27.267066 containerd[1899]: time="2026-03-02T12:49:27.267031899Z" level=info msg="StartContainer for \"fa285ca7266eb82cd1811a797c32d6ce21ff62b46af51bf68aea6f9577068720\"" Mar 2 12:49:27.268607 containerd[1899]: time="2026-03-02T12:49:27.268548911Z" level=info msg="connecting to shim fa285ca7266eb82cd1811a797c32d6ce21ff62b46af51bf68aea6f9577068720" address="unix:///run/containerd/s/4a7e3d152ae6384c296564c95881547677bacdc40e34c8766eafd26188fab70d" protocol=ttrpc version=3 Mar 2 12:49:27.284606 systemd[1]: Started cri-containerd-fa285ca7266eb82cd1811a797c32d6ce21ff62b46af51bf68aea6f9577068720.scope - libcontainer container fa285ca7266eb82cd1811a797c32d6ce21ff62b46af51bf68aea6f9577068720. Mar 2 12:49:27.307969 containerd[1899]: time="2026-03-02T12:49:27.307918030Z" level=info msg="connecting to shim 34297e654d3614fbb9bcbc83ffd2ab6072efe33a4f762d2fe7997948ba7cd4b8" address="unix:///run/containerd/s/252c4dc66bf883a359e76cf2d16fdda7a32f41f5185098683d3879b14167d66d" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:49:27.328740 systemd[1]: Started cri-containerd-34297e654d3614fbb9bcbc83ffd2ab6072efe33a4f762d2fe7997948ba7cd4b8.scope - libcontainer container 34297e654d3614fbb9bcbc83ffd2ab6072efe33a4f762d2fe7997948ba7cd4b8. Mar 2 12:49:27.350471 containerd[1899]: time="2026-03-02T12:49:27.350444825Z" level=info msg="StartContainer for \"fa285ca7266eb82cd1811a797c32d6ce21ff62b46af51bf68aea6f9577068720\" returns successfully" Mar 2 12:49:27.367900 containerd[1899]: time="2026-03-02T12:49:27.367816076Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6447996989-jxgk4,Uid:58fec222-c058-4252-86f6-3f6f395545c1,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"34297e654d3614fbb9bcbc83ffd2ab6072efe33a4f762d2fe7997948ba7cd4b8\"" Mar 2 12:49:27.370274 containerd[1899]: time="2026-03-02T12:49:27.370174125Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.3\"" Mar 2 12:49:28.770789 kubelet[3475]: I0302 12:49:28.770429 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-proxy-kk9tg" podStartSLOduration=2.770417894 podStartE2EDuration="2.770417894s" podCreationTimestamp="2026-03-02 12:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:49:27.430775772 +0000 UTC m=+8.133106551" watchObservedRunningTime="2026-03-02 12:49:28.770417894 +0000 UTC m=+9.472748633" Mar 2 12:49:28.918016 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2579426292.mount: Deactivated successfully. Mar 2 12:49:30.024942 containerd[1899]: time="2026-03-02T12:49:30.024900016Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:30.032017 containerd[1899]: time="2026-03-02T12:49:30.031886175Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.3: active requests=0, bytes read=25060789" Mar 2 12:49:30.035697 containerd[1899]: time="2026-03-02T12:49:30.035671369Z" level=info msg="ImageCreate event name:\"sha256:a94b0dfe779f8dc351e02e8988fd60aecb466000f13b6f00042ab83ebb237d87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:30.039890 containerd[1899]: time="2026-03-02T12:49:30.039848192Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:3b1a6762e1f3fae8490773b8f06ddd1e6775850febbece4d6002416f39adc670\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:30.040342 containerd[1899]: time="2026-03-02T12:49:30.040131546Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.3\" with image id \"sha256:a94b0dfe779f8dc351e02e8988fd60aecb466000f13b6f00042ab83ebb237d87\", repo tag \"quay.io/tigera/operator:v1.40.3\", repo digest \"quay.io/tigera/operator@sha256:3b1a6762e1f3fae8490773b8f06ddd1e6775850febbece4d6002416f39adc670\", size \"25056784\" in 2.669901603s" Mar 2 12:49:30.040342 containerd[1899]: time="2026-03-02T12:49:30.040155667Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.3\" returns image reference \"sha256:a94b0dfe779f8dc351e02e8988fd60aecb466000f13b6f00042ab83ebb237d87\"" Mar 2 12:49:30.048422 containerd[1899]: time="2026-03-02T12:49:30.048394022Z" level=info msg="CreateContainer within sandbox \"34297e654d3614fbb9bcbc83ffd2ab6072efe33a4f762d2fe7997948ba7cd4b8\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 2 12:49:30.066843 containerd[1899]: time="2026-03-02T12:49:30.066250490Z" level=info msg="Container d30d822a38601684dfa8fe6cc0f1659217fd38e12633e82dddecc8c6c6e7641d: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:30.069390 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3083732627.mount: Deactivated successfully. Mar 2 12:49:30.081660 containerd[1899]: time="2026-03-02T12:49:30.081632770Z" level=info msg="CreateContainer within sandbox \"34297e654d3614fbb9bcbc83ffd2ab6072efe33a4f762d2fe7997948ba7cd4b8\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"d30d822a38601684dfa8fe6cc0f1659217fd38e12633e82dddecc8c6c6e7641d\"" Mar 2 12:49:30.082120 containerd[1899]: time="2026-03-02T12:49:30.082063480Z" level=info msg="StartContainer for \"d30d822a38601684dfa8fe6cc0f1659217fd38e12633e82dddecc8c6c6e7641d\"" Mar 2 12:49:30.082695 containerd[1899]: time="2026-03-02T12:49:30.082669757Z" level=info msg="connecting to shim d30d822a38601684dfa8fe6cc0f1659217fd38e12633e82dddecc8c6c6e7641d" address="unix:///run/containerd/s/252c4dc66bf883a359e76cf2d16fdda7a32f41f5185098683d3879b14167d66d" protocol=ttrpc version=3 Mar 2 12:49:30.100598 systemd[1]: Started cri-containerd-d30d822a38601684dfa8fe6cc0f1659217fd38e12633e82dddecc8c6c6e7641d.scope - libcontainer container d30d822a38601684dfa8fe6cc0f1659217fd38e12633e82dddecc8c6c6e7641d. Mar 2 12:49:30.124633 containerd[1899]: time="2026-03-02T12:49:30.124588867Z" level=info msg="StartContainer for \"d30d822a38601684dfa8fe6cc0f1659217fd38e12633e82dddecc8c6c6e7641d\" returns successfully" Mar 2 12:49:34.743120 kubelet[3475]: I0302 12:49:34.742856 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6447996989-jxgk4" podStartSLOduration=6.071522028 podStartE2EDuration="8.742844143s" podCreationTimestamp="2026-03-02 12:49:26 +0000 UTC" firstStartedPulling="2026-03-02 12:49:27.369576105 +0000 UTC m=+8.071906844" lastFinishedPulling="2026-03-02 12:49:30.04089822 +0000 UTC m=+10.743228959" observedRunningTime="2026-03-02 12:49:30.435540086 +0000 UTC m=+11.137870857" watchObservedRunningTime="2026-03-02 12:49:34.742844143 +0000 UTC m=+15.445174882" Mar 2 12:49:35.062012 sudo[2374]: pam_unix(sudo:session): session closed for user root Mar 2 12:49:35.141940 sshd[2373]: Connection closed by 10.200.16.10 port 52260 Mar 2 12:49:35.142652 sshd-session[2370]: pam_unix(sshd:session): session closed for user core Mar 2 12:49:35.146898 systemd[1]: sshd@6-10.200.20.38:22-10.200.16.10:52260.service: Deactivated successfully. Mar 2 12:49:35.149339 systemd[1]: session-9.scope: Deactivated successfully. Mar 2 12:49:35.149912 systemd[1]: session-9.scope: Consumed 2.238s CPU time, 218.6M memory peak. Mar 2 12:49:35.152580 systemd-logind[1870]: Session 9 logged out. Waiting for processes to exit. Mar 2 12:49:35.155396 systemd-logind[1870]: Removed session 9. Mar 2 12:49:39.486761 systemd[1]: Created slice kubepods-besteffort-pod2c68d25b_46ae_424d_bf9b_5c3c8c96edf6.slice - libcontainer container kubepods-besteffort-pod2c68d25b_46ae_424d_bf9b_5c3c8c96edf6.slice. Mar 2 12:49:39.495956 kubelet[3475]: I0302 12:49:39.495923 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4czl\" (UniqueName: \"kubernetes.io/projected/2c68d25b-46ae-424d-bf9b-5c3c8c96edf6-kube-api-access-q4czl\") pod \"calico-typha-744bcc57c6-g97j9\" (UID: \"2c68d25b-46ae-424d-bf9b-5c3c8c96edf6\") " pod="calico-system/calico-typha-744bcc57c6-g97j9" Mar 2 12:49:39.495956 kubelet[3475]: I0302 12:49:39.495953 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c68d25b-46ae-424d-bf9b-5c3c8c96edf6-tigera-ca-bundle\") pod \"calico-typha-744bcc57c6-g97j9\" (UID: \"2c68d25b-46ae-424d-bf9b-5c3c8c96edf6\") " pod="calico-system/calico-typha-744bcc57c6-g97j9" Mar 2 12:49:39.496200 kubelet[3475]: I0302 12:49:39.495964 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/2c68d25b-46ae-424d-bf9b-5c3c8c96edf6-typha-certs\") pod \"calico-typha-744bcc57c6-g97j9\" (UID: \"2c68d25b-46ae-424d-bf9b-5c3c8c96edf6\") " pod="calico-system/calico-typha-744bcc57c6-g97j9" Mar 2 12:49:39.591318 systemd[1]: Created slice kubepods-besteffort-pod511139ce_86d9_47bd_9fcd_e7535dd8a0c3.slice - libcontainer container kubepods-besteffort-pod511139ce_86d9_47bd_9fcd_e7535dd8a0c3.slice. Mar 2 12:49:39.596907 kubelet[3475]: I0302 12:49:39.596875 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-bpffs\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.596907 kubelet[3475]: I0302 12:49:39.596907 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-flexvol-driver-host\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597001 kubelet[3475]: I0302 12:49:39.596918 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-node-certs\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597001 kubelet[3475]: I0302 12:49:39.596935 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-cni-net-dir\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597001 kubelet[3475]: I0302 12:49:39.596946 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-cni-bin-dir\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597001 kubelet[3475]: I0302 12:49:39.596957 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-xtables-lock\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597001 kubelet[3475]: I0302 12:49:39.596967 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-lib-modules\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597075 kubelet[3475]: I0302 12:49:39.596976 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm65l\" (UniqueName: \"kubernetes.io/projected/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-kube-api-access-pm65l\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597075 kubelet[3475]: I0302 12:49:39.596984 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-var-lib-calico\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597075 kubelet[3475]: I0302 12:49:39.597005 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-nodeproc\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597075 kubelet[3475]: I0302 12:49:39.597014 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-sys-fs\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597075 kubelet[3475]: I0302 12:49:39.597024 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-var-run-calico\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597151 kubelet[3475]: I0302 12:49:39.597032 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-cni-log-dir\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597151 kubelet[3475]: I0302 12:49:39.597041 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-tigera-ca-bundle\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.597151 kubelet[3475]: I0302 12:49:39.597052 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/511139ce-86d9-47bd-9fcd-e7535dd8a0c3-policysync\") pod \"calico-node-h5k9w\" (UID: \"511139ce-86d9-47bd-9fcd-e7535dd8a0c3\") " pod="calico-system/calico-node-h5k9w" Mar 2 12:49:39.702870 kubelet[3475]: E0302 12:49:39.702832 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.702870 kubelet[3475]: W0302 12:49:39.702852 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.703574 kubelet[3475]: E0302 12:49:39.702940 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.703723 kubelet[3475]: E0302 12:49:39.703692 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.704240 kubelet[3475]: W0302 12:49:39.704223 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.704628 kubelet[3475]: E0302 12:49:39.704305 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.705611 kubelet[3475]: E0302 12:49:39.705591 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.705611 kubelet[3475]: W0302 12:49:39.705606 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.705762 kubelet[3475]: E0302 12:49:39.705618 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.706510 kubelet[3475]: E0302 12:49:39.706195 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:39.707558 kubelet[3475]: E0302 12:49:39.707041 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.707835 kubelet[3475]: W0302 12:49:39.707631 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.707835 kubelet[3475]: E0302 12:49:39.707649 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.709045 kubelet[3475]: E0302 12:49:39.708789 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.709045 kubelet[3475]: W0302 12:49:39.708800 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.709045 kubelet[3475]: E0302 12:49:39.708811 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.709662 kubelet[3475]: E0302 12:49:39.709255 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.709662 kubelet[3475]: W0302 12:49:39.709267 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.709662 kubelet[3475]: E0302 12:49:39.709279 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.709662 kubelet[3475]: E0302 12:49:39.709508 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.709662 kubelet[3475]: W0302 12:49:39.709519 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.709662 kubelet[3475]: E0302 12:49:39.709528 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.710578 kubelet[3475]: E0302 12:49:39.710564 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.710663 kubelet[3475]: W0302 12:49:39.710653 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.710735 kubelet[3475]: E0302 12:49:39.710724 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.710965 kubelet[3475]: E0302 12:49:39.710954 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.711112 kubelet[3475]: W0302 12:49:39.711007 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.711112 kubelet[3475]: E0302 12:49:39.711020 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.711346 kubelet[3475]: E0302 12:49:39.711334 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.711565 kubelet[3475]: W0302 12:49:39.711408 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.711565 kubelet[3475]: E0302 12:49:39.711421 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.711624 kubelet[3475]: E0302 12:49:39.711610 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.711624 kubelet[3475]: W0302 12:49:39.711621 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.711624 kubelet[3475]: E0302 12:49:39.711632 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.712370 kubelet[3475]: E0302 12:49:39.712349 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.712370 kubelet[3475]: W0302 12:49:39.712362 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.712370 kubelet[3475]: E0302 12:49:39.712373 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.714430 kubelet[3475]: E0302 12:49:39.714414 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.714672 kubelet[3475]: W0302 12:49:39.714597 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.714672 kubelet[3475]: E0302 12:49:39.714613 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.715150 kubelet[3475]: E0302 12:49:39.715130 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.715150 kubelet[3475]: W0302 12:49:39.715146 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.715358 kubelet[3475]: E0302 12:49:39.715157 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.715538 kubelet[3475]: E0302 12:49:39.715521 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.715538 kubelet[3475]: W0302 12:49:39.715533 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.715647 kubelet[3475]: E0302 12:49:39.715543 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.715666 kubelet[3475]: E0302 12:49:39.715652 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.715666 kubelet[3475]: W0302 12:49:39.715657 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.715666 kubelet[3475]: E0302 12:49:39.715664 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.715798 kubelet[3475]: E0302 12:49:39.715771 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.715798 kubelet[3475]: W0302 12:49:39.715781 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.715798 kubelet[3475]: E0302 12:49:39.715787 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.715925 kubelet[3475]: E0302 12:49:39.715898 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.715925 kubelet[3475]: W0302 12:49:39.715903 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.715925 kubelet[3475]: E0302 12:49:39.715909 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716039 kubelet[3475]: E0302 12:49:39.716015 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716039 kubelet[3475]: W0302 12:49:39.716023 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716039 kubelet[3475]: E0302 12:49:39.716030 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716217 kubelet[3475]: E0302 12:49:39.716127 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716217 kubelet[3475]: W0302 12:49:39.716132 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716217 kubelet[3475]: E0302 12:49:39.716138 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716281 kubelet[3475]: E0302 12:49:39.716226 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716281 kubelet[3475]: W0302 12:49:39.716230 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716281 kubelet[3475]: E0302 12:49:39.716236 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716428 kubelet[3475]: E0302 12:49:39.716328 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716428 kubelet[3475]: W0302 12:49:39.716333 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716428 kubelet[3475]: E0302 12:49:39.716338 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716428 kubelet[3475]: E0302 12:49:39.716418 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716428 kubelet[3475]: W0302 12:49:39.716422 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716428 kubelet[3475]: E0302 12:49:39.716427 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716569 kubelet[3475]: E0302 12:49:39.716555 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716569 kubelet[3475]: W0302 12:49:39.716562 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716569 kubelet[3475]: E0302 12:49:39.716569 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716726 kubelet[3475]: E0302 12:49:39.716716 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716726 kubelet[3475]: W0302 12:49:39.716722 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716726 kubelet[3475]: E0302 12:49:39.716729 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716852 kubelet[3475]: E0302 12:49:39.716838 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716852 kubelet[3475]: W0302 12:49:39.716847 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716852 kubelet[3475]: E0302 12:49:39.716854 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.716985 kubelet[3475]: E0302 12:49:39.716967 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.716985 kubelet[3475]: W0302 12:49:39.716975 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.716985 kubelet[3475]: E0302 12:49:39.716982 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.717099 kubelet[3475]: E0302 12:49:39.717087 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.717099 kubelet[3475]: W0302 12:49:39.717095 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.717155 kubelet[3475]: E0302 12:49:39.717100 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.717220 kubelet[3475]: E0302 12:49:39.717207 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.717220 kubelet[3475]: W0302 12:49:39.717215 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.717220 kubelet[3475]: E0302 12:49:39.717220 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.717411 kubelet[3475]: E0302 12:49:39.717350 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.717411 kubelet[3475]: W0302 12:49:39.717358 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.717411 kubelet[3475]: E0302 12:49:39.717364 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.717979 kubelet[3475]: E0302 12:49:39.717959 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.717979 kubelet[3475]: W0302 12:49:39.717975 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.718210 kubelet[3475]: E0302 12:49:39.717995 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.718289 kubelet[3475]: E0302 12:49:39.718271 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.718289 kubelet[3475]: W0302 12:49:39.718286 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.718599 kubelet[3475]: E0302 12:49:39.718295 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.718772 kubelet[3475]: E0302 12:49:39.718738 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.718772 kubelet[3475]: W0302 12:49:39.718753 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.718772 kubelet[3475]: E0302 12:49:39.718763 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.719184 kubelet[3475]: E0302 12:49:39.719151 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.719184 kubelet[3475]: W0302 12:49:39.719182 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.719230 kubelet[3475]: E0302 12:49:39.719193 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.719462 kubelet[3475]: E0302 12:49:39.719448 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.719462 kubelet[3475]: W0302 12:49:39.719461 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.719523 kubelet[3475]: E0302 12:49:39.719482 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.719755 kubelet[3475]: E0302 12:49:39.719738 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.719755 kubelet[3475]: W0302 12:49:39.719751 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.719819 kubelet[3475]: E0302 12:49:39.719760 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.720103 kubelet[3475]: E0302 12:49:39.720085 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.720103 kubelet[3475]: W0302 12:49:39.720100 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.720155 kubelet[3475]: E0302 12:49:39.720109 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.720329 kubelet[3475]: E0302 12:49:39.720281 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.720329 kubelet[3475]: W0302 12:49:39.720320 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.720329 kubelet[3475]: E0302 12:49:39.720331 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.720540 kubelet[3475]: E0302 12:49:39.720525 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.720540 kubelet[3475]: W0302 12:49:39.720538 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.720593 kubelet[3475]: E0302 12:49:39.720548 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.720826 kubelet[3475]: E0302 12:49:39.720805 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.720882 kubelet[3475]: W0302 12:49:39.720843 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.720882 kubelet[3475]: E0302 12:49:39.720856 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.721129 kubelet[3475]: E0302 12:49:39.721114 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.721129 kubelet[3475]: W0302 12:49:39.721127 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.721184 kubelet[3475]: E0302 12:49:39.721153 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.728774 kubelet[3475]: E0302 12:49:39.728755 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.728774 kubelet[3475]: W0302 12:49:39.728770 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.728858 kubelet[3475]: E0302 12:49:39.728780 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.796023 kubelet[3475]: E0302 12:49:39.795467 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.796023 kubelet[3475]: W0302 12:49:39.795482 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.796023 kubelet[3475]: E0302 12:49:39.795535 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.796023 kubelet[3475]: E0302 12:49:39.795670 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.796023 kubelet[3475]: W0302 12:49:39.795677 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.796023 kubelet[3475]: E0302 12:49:39.795685 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.796023 kubelet[3475]: E0302 12:49:39.795798 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.796023 kubelet[3475]: W0302 12:49:39.795804 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.796023 kubelet[3475]: E0302 12:49:39.795811 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.796023 kubelet[3475]: E0302 12:49:39.795922 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.796213 kubelet[3475]: W0302 12:49:39.795928 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.796213 kubelet[3475]: E0302 12:49:39.795934 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.796213 kubelet[3475]: E0302 12:49:39.796091 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.796213 kubelet[3475]: W0302 12:49:39.796098 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.796213 kubelet[3475]: E0302 12:49:39.796106 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.797101 kubelet[3475]: E0302 12:49:39.797083 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.797101 kubelet[3475]: W0302 12:49:39.797096 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.797159 kubelet[3475]: E0302 12:49:39.797105 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.797242 kubelet[3475]: E0302 12:49:39.797224 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.797242 kubelet[3475]: W0302 12:49:39.797234 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.797242 kubelet[3475]: E0302 12:49:39.797241 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.797364 kubelet[3475]: E0302 12:49:39.797350 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.797364 kubelet[3475]: W0302 12:49:39.797358 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.797364 kubelet[3475]: E0302 12:49:39.797364 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.797523 kubelet[3475]: E0302 12:49:39.797510 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.797523 kubelet[3475]: W0302 12:49:39.797519 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.797578 kubelet[3475]: E0302 12:49:39.797526 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.797666 kubelet[3475]: E0302 12:49:39.797654 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.797666 kubelet[3475]: W0302 12:49:39.797662 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.797666 kubelet[3475]: E0302 12:49:39.797668 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.797854 kubelet[3475]: E0302 12:49:39.797840 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.797854 kubelet[3475]: W0302 12:49:39.797849 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.797902 kubelet[3475]: E0302 12:49:39.797857 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.798834 containerd[1899]: time="2026-03-02T12:49:39.798218967Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-744bcc57c6-g97j9,Uid:2c68d25b-46ae-424d-bf9b-5c3c8c96edf6,Namespace:calico-system,Attempt:0,}" Mar 2 12:49:39.799061 kubelet[3475]: E0302 12:49:39.798263 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.799061 kubelet[3475]: W0302 12:49:39.798272 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.799061 kubelet[3475]: E0302 12:49:39.798290 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.799061 kubelet[3475]: E0302 12:49:39.798521 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.799061 kubelet[3475]: W0302 12:49:39.798530 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.799061 kubelet[3475]: E0302 12:49:39.798539 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.799061 kubelet[3475]: E0302 12:49:39.798673 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.799061 kubelet[3475]: W0302 12:49:39.798681 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.799061 kubelet[3475]: E0302 12:49:39.798689 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.799387 kubelet[3475]: E0302 12:49:39.799281 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.799387 kubelet[3475]: W0302 12:49:39.799292 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.799387 kubelet[3475]: E0302 12:49:39.799308 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.799605 kubelet[3475]: E0302 12:49:39.799594 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.799695 kubelet[3475]: W0302 12:49:39.799651 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.799695 kubelet[3475]: E0302 12:49:39.799678 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.800029 kubelet[3475]: E0302 12:49:39.800018 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.800172 kubelet[3475]: W0302 12:49:39.800061 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.800172 kubelet[3475]: E0302 12:49:39.800074 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.800459 kubelet[3475]: E0302 12:49:39.800407 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.800459 kubelet[3475]: W0302 12:49:39.800419 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.800459 kubelet[3475]: E0302 12:49:39.800428 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.800857 kubelet[3475]: E0302 12:49:39.800801 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.800857 kubelet[3475]: W0302 12:49:39.800810 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.800857 kubelet[3475]: E0302 12:49:39.800819 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.801091 kubelet[3475]: E0302 12:49:39.801081 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.801163 kubelet[3475]: W0302 12:49:39.801140 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.801218 kubelet[3475]: E0302 12:49:39.801206 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.801576 kubelet[3475]: E0302 12:49:39.801564 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.801716 kubelet[3475]: W0302 12:49:39.801639 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.801786 kubelet[3475]: E0302 12:49:39.801775 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.801901 kubelet[3475]: I0302 12:49:39.801853 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/07af8044-862a-4c44-bd2c-12b2e3c44607-registration-dir\") pod \"csi-node-driver-wtzgs\" (UID: \"07af8044-862a-4c44-bd2c-12b2e3c44607\") " pod="calico-system/csi-node-driver-wtzgs" Mar 2 12:49:39.802133 kubelet[3475]: E0302 12:49:39.802117 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.802267 kubelet[3475]: W0302 12:49:39.802187 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.802267 kubelet[3475]: E0302 12:49:39.802201 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.802267 kubelet[3475]: I0302 12:49:39.802227 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07af8044-862a-4c44-bd2c-12b2e3c44607-kubelet-dir\") pod \"csi-node-driver-wtzgs\" (UID: \"07af8044-862a-4c44-bd2c-12b2e3c44607\") " pod="calico-system/csi-node-driver-wtzgs" Mar 2 12:49:39.802525 kubelet[3475]: E0302 12:49:39.802479 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.802630 kubelet[3475]: W0302 12:49:39.802586 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.802630 kubelet[3475]: E0302 12:49:39.802603 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.802798 kubelet[3475]: I0302 12:49:39.802729 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/07af8044-862a-4c44-bd2c-12b2e3c44607-socket-dir\") pod \"csi-node-driver-wtzgs\" (UID: \"07af8044-862a-4c44-bd2c-12b2e3c44607\") " pod="calico-system/csi-node-driver-wtzgs" Mar 2 12:49:39.802984 kubelet[3475]: E0302 12:49:39.802952 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.802984 kubelet[3475]: W0302 12:49:39.802962 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.803095 kubelet[3475]: E0302 12:49:39.803049 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.803258 kubelet[3475]: E0302 12:49:39.803248 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.803339 kubelet[3475]: W0302 12:49:39.803316 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.803339 kubelet[3475]: E0302 12:49:39.803330 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.803698 kubelet[3475]: E0302 12:49:39.803669 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.803698 kubelet[3475]: W0302 12:49:39.803680 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.803698 kubelet[3475]: E0302 12:49:39.803689 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.803966 kubelet[3475]: E0302 12:49:39.803938 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.803966 kubelet[3475]: W0302 12:49:39.803949 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.803966 kubelet[3475]: E0302 12:49:39.803957 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.804296 kubelet[3475]: E0302 12:49:39.804267 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.804296 kubelet[3475]: W0302 12:49:39.804278 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.804296 kubelet[3475]: E0302 12:49:39.804286 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.804468 kubelet[3475]: I0302 12:49:39.804445 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/07af8044-862a-4c44-bd2c-12b2e3c44607-varrun\") pod \"csi-node-driver-wtzgs\" (UID: \"07af8044-862a-4c44-bd2c-12b2e3c44607\") " pod="calico-system/csi-node-driver-wtzgs" Mar 2 12:49:39.804669 kubelet[3475]: E0302 12:49:39.804641 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.804669 kubelet[3475]: W0302 12:49:39.804651 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.804669 kubelet[3475]: E0302 12:49:39.804659 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.804932 kubelet[3475]: E0302 12:49:39.804903 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.804932 kubelet[3475]: W0302 12:49:39.804914 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.804932 kubelet[3475]: E0302 12:49:39.804923 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.805203 kubelet[3475]: E0302 12:49:39.805185 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.805203 kubelet[3475]: W0302 12:49:39.805199 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.805276 kubelet[3475]: E0302 12:49:39.805211 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.805276 kubelet[3475]: I0302 12:49:39.805230 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d6vj\" (UniqueName: \"kubernetes.io/projected/07af8044-862a-4c44-bd2c-12b2e3c44607-kube-api-access-7d6vj\") pod \"csi-node-driver-wtzgs\" (UID: \"07af8044-862a-4c44-bd2c-12b2e3c44607\") " pod="calico-system/csi-node-driver-wtzgs" Mar 2 12:49:39.805378 kubelet[3475]: E0302 12:49:39.805365 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.805378 kubelet[3475]: W0302 12:49:39.805375 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.805425 kubelet[3475]: E0302 12:49:39.805383 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.805857 kubelet[3475]: E0302 12:49:39.805840 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.805857 kubelet[3475]: W0302 12:49:39.805854 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.805949 kubelet[3475]: E0302 12:49:39.805864 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.805996 kubelet[3475]: E0302 12:49:39.805984 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.805996 kubelet[3475]: W0302 12:49:39.805992 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.806046 kubelet[3475]: E0302 12:49:39.805999 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.806090 kubelet[3475]: E0302 12:49:39.806081 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.806090 kubelet[3475]: W0302 12:49:39.806087 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.806126 kubelet[3475]: E0302 12:49:39.806092 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.848578 containerd[1899]: time="2026-03-02T12:49:39.848533302Z" level=info msg="connecting to shim 3df209cf8c84a5fae29e24ab2d93cad16b9f2942112bc586ca965880a80476ee" address="unix:///run/containerd/s/3438c7eec0b9a37b2a6a9a630f472643f46a3609fcd05a982f1633ef027fc2c9" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:49:39.870621 systemd[1]: Started cri-containerd-3df209cf8c84a5fae29e24ab2d93cad16b9f2942112bc586ca965880a80476ee.scope - libcontainer container 3df209cf8c84a5fae29e24ab2d93cad16b9f2942112bc586ca965880a80476ee. Mar 2 12:49:39.898018 containerd[1899]: time="2026-03-02T12:49:39.897956568Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-744bcc57c6-g97j9,Uid:2c68d25b-46ae-424d-bf9b-5c3c8c96edf6,Namespace:calico-system,Attempt:0,} returns sandbox id \"3df209cf8c84a5fae29e24ab2d93cad16b9f2942112bc586ca965880a80476ee\"" Mar 2 12:49:39.901544 containerd[1899]: time="2026-03-02T12:49:39.901513617Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.3\"" Mar 2 12:49:39.902342 containerd[1899]: time="2026-03-02T12:49:39.902318556Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-h5k9w,Uid:511139ce-86d9-47bd-9fcd-e7535dd8a0c3,Namespace:calico-system,Attempt:0,}" Mar 2 12:49:39.906161 kubelet[3475]: E0302 12:49:39.906138 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.906161 kubelet[3475]: W0302 12:49:39.906156 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.906253 kubelet[3475]: E0302 12:49:39.906170 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.906357 kubelet[3475]: E0302 12:49:39.906332 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.906357 kubelet[3475]: W0302 12:49:39.906344 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.906357 kubelet[3475]: E0302 12:49:39.906352 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.906561 kubelet[3475]: E0302 12:49:39.906544 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.906561 kubelet[3475]: W0302 12:49:39.906557 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.906623 kubelet[3475]: E0302 12:49:39.906565 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.906754 kubelet[3475]: E0302 12:49:39.906739 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.906754 kubelet[3475]: W0302 12:49:39.906751 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.906808 kubelet[3475]: E0302 12:49:39.906759 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.906932 kubelet[3475]: E0302 12:49:39.906918 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.906932 kubelet[3475]: W0302 12:49:39.906928 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.906988 kubelet[3475]: E0302 12:49:39.906936 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.907121 kubelet[3475]: E0302 12:49:39.907107 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.907121 kubelet[3475]: W0302 12:49:39.907117 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.907178 kubelet[3475]: E0302 12:49:39.907124 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.907363 kubelet[3475]: E0302 12:49:39.907350 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.907363 kubelet[3475]: W0302 12:49:39.907361 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.907420 kubelet[3475]: E0302 12:49:39.907370 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.907544 kubelet[3475]: E0302 12:49:39.907529 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.907544 kubelet[3475]: W0302 12:49:39.907540 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.907583 kubelet[3475]: E0302 12:49:39.907549 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909180 kubelet[3475]: E0302 12:49:39.907679 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909180 kubelet[3475]: W0302 12:49:39.907685 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909180 kubelet[3475]: E0302 12:49:39.907692 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909180 kubelet[3475]: E0302 12:49:39.907811 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909180 kubelet[3475]: W0302 12:49:39.907817 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909180 kubelet[3475]: E0302 12:49:39.907823 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909180 kubelet[3475]: E0302 12:49:39.907939 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909180 kubelet[3475]: W0302 12:49:39.907945 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909180 kubelet[3475]: E0302 12:49:39.907951 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909180 kubelet[3475]: E0302 12:49:39.908262 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909349 kubelet[3475]: W0302 12:49:39.908273 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909349 kubelet[3475]: E0302 12:49:39.908282 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909349 kubelet[3475]: E0302 12:49:39.908441 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909349 kubelet[3475]: W0302 12:49:39.908447 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909349 kubelet[3475]: E0302 12:49:39.908453 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909349 kubelet[3475]: E0302 12:49:39.908623 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909349 kubelet[3475]: W0302 12:49:39.908630 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909349 kubelet[3475]: E0302 12:49:39.908637 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909349 kubelet[3475]: E0302 12:49:39.908750 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909349 kubelet[3475]: W0302 12:49:39.908755 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909477 kubelet[3475]: E0302 12:49:39.908761 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909477 kubelet[3475]: E0302 12:49:39.908952 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909477 kubelet[3475]: W0302 12:49:39.908961 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909477 kubelet[3475]: E0302 12:49:39.908968 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909477 kubelet[3475]: E0302 12:49:39.909070 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909477 kubelet[3475]: W0302 12:49:39.909076 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909477 kubelet[3475]: E0302 12:49:39.909083 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909688 kubelet[3475]: E0302 12:49:39.909594 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909688 kubelet[3475]: W0302 12:49:39.909604 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909688 kubelet[3475]: E0302 12:49:39.909613 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909766 kubelet[3475]: E0302 12:49:39.909740 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909766 kubelet[3475]: W0302 12:49:39.909752 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909766 kubelet[3475]: E0302 12:49:39.909759 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.909916 kubelet[3475]: E0302 12:49:39.909902 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.909916 kubelet[3475]: W0302 12:49:39.909911 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.909973 kubelet[3475]: E0302 12:49:39.909918 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.910064 kubelet[3475]: E0302 12:49:39.910049 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.910064 kubelet[3475]: W0302 12:49:39.910060 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.910112 kubelet[3475]: E0302 12:49:39.910066 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.910193 kubelet[3475]: E0302 12:49:39.910178 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.910193 kubelet[3475]: W0302 12:49:39.910188 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.910244 kubelet[3475]: E0302 12:49:39.910194 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.910323 kubelet[3475]: E0302 12:49:39.910309 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.910323 kubelet[3475]: W0302 12:49:39.910319 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.910373 kubelet[3475]: E0302 12:49:39.910325 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.910535 kubelet[3475]: E0302 12:49:39.910519 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.910535 kubelet[3475]: W0302 12:49:39.910530 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.910580 kubelet[3475]: E0302 12:49:39.910537 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.911015 kubelet[3475]: E0302 12:49:39.910993 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.911015 kubelet[3475]: W0302 12:49:39.911009 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.911103 kubelet[3475]: E0302 12:49:39.911020 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.919042 kubelet[3475]: E0302 12:49:39.919020 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:39.919042 kubelet[3475]: W0302 12:49:39.919035 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:39.919042 kubelet[3475]: E0302 12:49:39.919046 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:39.955767 containerd[1899]: time="2026-03-02T12:49:39.955730981Z" level=info msg="connecting to shim 5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539" address="unix:///run/containerd/s/2b27b6f4cf4c8b74067644f9280c1a5e82cb184a0c3d4e011ad8910a9c617a92" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:49:39.972658 systemd[1]: Started cri-containerd-5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539.scope - libcontainer container 5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539. Mar 2 12:49:39.992021 containerd[1899]: time="2026-03-02T12:49:39.991991087Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-h5k9w,Uid:511139ce-86d9-47bd-9fcd-e7535dd8a0c3,Namespace:calico-system,Attempt:0,} returns sandbox id \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\"" Mar 2 12:49:41.381448 kubelet[3475]: E0302 12:49:41.380844 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:41.395384 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount698954684.mount: Deactivated successfully. Mar 2 12:49:42.353899 containerd[1899]: time="2026-03-02T12:49:42.353853463Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:42.357310 containerd[1899]: time="2026-03-02T12:49:42.357283252Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.3: active requests=0, bytes read=33841852" Mar 2 12:49:42.360763 containerd[1899]: time="2026-03-02T12:49:42.360737905Z" level=info msg="ImageCreate event name:\"sha256:d28a261c14ff1c1c526940695055ffc414471b39d275a706eac99ccbbd5fdc62\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:42.364976 containerd[1899]: time="2026-03-02T12:49:42.364949569Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:3e62cf98a20c42a1786397d0192cfb639634ef95c6f463ab92f0439a5c1a4ae5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:42.365396 containerd[1899]: time="2026-03-02T12:49:42.365373191Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.3\" with image id \"sha256:d28a261c14ff1c1c526940695055ffc414471b39d275a706eac99ccbbd5fdc62\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:3e62cf98a20c42a1786397d0192cfb639634ef95c6f463ab92f0439a5c1a4ae5\", size \"33841706\" in 2.463825429s" Mar 2 12:49:42.365417 containerd[1899]: time="2026-03-02T12:49:42.365400328Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.3\" returns image reference \"sha256:d28a261c14ff1c1c526940695055ffc414471b39d275a706eac99ccbbd5fdc62\"" Mar 2 12:49:42.366629 containerd[1899]: time="2026-03-02T12:49:42.366599809Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\"" Mar 2 12:49:42.384540 containerd[1899]: time="2026-03-02T12:49:42.384516378Z" level=info msg="CreateContainer within sandbox \"3df209cf8c84a5fae29e24ab2d93cad16b9f2942112bc586ca965880a80476ee\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 2 12:49:42.415098 containerd[1899]: time="2026-03-02T12:49:42.414572849Z" level=info msg="Container a523e3bfbdf0dc72043da14ade3e527efeedc63f0a5f37aaa7b3c9ab9be55cac: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:42.417012 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3638328700.mount: Deactivated successfully. Mar 2 12:49:42.434774 containerd[1899]: time="2026-03-02T12:49:42.434745143Z" level=info msg="CreateContainer within sandbox \"3df209cf8c84a5fae29e24ab2d93cad16b9f2942112bc586ca965880a80476ee\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"a523e3bfbdf0dc72043da14ade3e527efeedc63f0a5f37aaa7b3c9ab9be55cac\"" Mar 2 12:49:42.435855 containerd[1899]: time="2026-03-02T12:49:42.435830140Z" level=info msg="StartContainer for \"a523e3bfbdf0dc72043da14ade3e527efeedc63f0a5f37aaa7b3c9ab9be55cac\"" Mar 2 12:49:42.436709 containerd[1899]: time="2026-03-02T12:49:42.436622511Z" level=info msg="connecting to shim a523e3bfbdf0dc72043da14ade3e527efeedc63f0a5f37aaa7b3c9ab9be55cac" address="unix:///run/containerd/s/3438c7eec0b9a37b2a6a9a630f472643f46a3609fcd05a982f1633ef027fc2c9" protocol=ttrpc version=3 Mar 2 12:49:42.451609 systemd[1]: Started cri-containerd-a523e3bfbdf0dc72043da14ade3e527efeedc63f0a5f37aaa7b3c9ab9be55cac.scope - libcontainer container a523e3bfbdf0dc72043da14ade3e527efeedc63f0a5f37aaa7b3c9ab9be55cac. Mar 2 12:49:42.485228 containerd[1899]: time="2026-03-02T12:49:42.485198780Z" level=info msg="StartContainer for \"a523e3bfbdf0dc72043da14ade3e527efeedc63f0a5f37aaa7b3c9ab9be55cac\" returns successfully" Mar 2 12:49:43.381462 kubelet[3475]: E0302 12:49:43.381133 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:43.526569 kubelet[3475]: E0302 12:49:43.526544 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.526696 kubelet[3475]: W0302 12:49:43.526564 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.526696 kubelet[3475]: E0302 12:49:43.526614 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.526768 kubelet[3475]: E0302 12:49:43.526758 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.526768 kubelet[3475]: W0302 12:49:43.526766 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.526837 kubelet[3475]: E0302 12:49:43.526774 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.526894 kubelet[3475]: E0302 12:49:43.526870 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.526894 kubelet[3475]: W0302 12:49:43.526886 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.526894 kubelet[3475]: E0302 12:49:43.526892 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.526988 kubelet[3475]: E0302 12:49:43.526984 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527006 kubelet[3475]: W0302 12:49:43.526989 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527006 kubelet[3475]: E0302 12:49:43.526995 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527107 kubelet[3475]: E0302 12:49:43.527093 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527107 kubelet[3475]: W0302 12:49:43.527107 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527160 kubelet[3475]: E0302 12:49:43.527114 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527212 kubelet[3475]: E0302 12:49:43.527203 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527212 kubelet[3475]: W0302 12:49:43.527210 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527255 kubelet[3475]: E0302 12:49:43.527216 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527317 kubelet[3475]: E0302 12:49:43.527306 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527317 kubelet[3475]: W0302 12:49:43.527313 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527360 kubelet[3475]: E0302 12:49:43.527318 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527420 kubelet[3475]: E0302 12:49:43.527409 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527420 kubelet[3475]: W0302 12:49:43.527416 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527462 kubelet[3475]: E0302 12:49:43.527422 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527535 kubelet[3475]: E0302 12:49:43.527524 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527535 kubelet[3475]: W0302 12:49:43.527531 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527589 kubelet[3475]: E0302 12:49:43.527537 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527633 kubelet[3475]: E0302 12:49:43.527618 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527633 kubelet[3475]: W0302 12:49:43.527631 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527690 kubelet[3475]: E0302 12:49:43.527636 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527717 kubelet[3475]: E0302 12:49:43.527713 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527737 kubelet[3475]: W0302 12:49:43.527718 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527737 kubelet[3475]: E0302 12:49:43.527723 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527818 kubelet[3475]: E0302 12:49:43.527809 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527818 kubelet[3475]: W0302 12:49:43.527815 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527860 kubelet[3475]: E0302 12:49:43.527820 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.527919 kubelet[3475]: E0302 12:49:43.527909 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.527919 kubelet[3475]: W0302 12:49:43.527916 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.527961 kubelet[3475]: E0302 12:49:43.527931 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.528028 kubelet[3475]: E0302 12:49:43.528020 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.528028 kubelet[3475]: W0302 12:49:43.528026 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.528077 kubelet[3475]: E0302 12:49:43.528031 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.528123 kubelet[3475]: E0302 12:49:43.528113 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.528123 kubelet[3475]: W0302 12:49:43.528119 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.528163 kubelet[3475]: E0302 12:49:43.528126 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.530499 kubelet[3475]: E0302 12:49:43.530455 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.530499 kubelet[3475]: W0302 12:49:43.530467 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.530499 kubelet[3475]: E0302 12:49:43.530478 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.530756 kubelet[3475]: E0302 12:49:43.530736 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.530756 kubelet[3475]: W0302 12:49:43.530751 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.530814 kubelet[3475]: E0302 12:49:43.530761 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.530904 kubelet[3475]: E0302 12:49:43.530893 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.530904 kubelet[3475]: W0302 12:49:43.530901 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.530945 kubelet[3475]: E0302 12:49:43.530909 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.531019 kubelet[3475]: E0302 12:49:43.531010 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.531019 kubelet[3475]: W0302 12:49:43.531017 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.531056 kubelet[3475]: E0302 12:49:43.531023 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.531180 kubelet[3475]: E0302 12:49:43.531167 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.531180 kubelet[3475]: W0302 12:49:43.531176 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.531222 kubelet[3475]: E0302 12:49:43.531186 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.531549 kubelet[3475]: E0302 12:49:43.531533 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.531549 kubelet[3475]: W0302 12:49:43.531547 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.531598 kubelet[3475]: E0302 12:49:43.531555 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.531739 kubelet[3475]: E0302 12:49:43.531725 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.531739 kubelet[3475]: W0302 12:49:43.531735 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.531788 kubelet[3475]: E0302 12:49:43.531742 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.531915 kubelet[3475]: E0302 12:49:43.531902 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.531915 kubelet[3475]: W0302 12:49:43.531912 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.531966 kubelet[3475]: E0302 12:49:43.531919 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.532054 kubelet[3475]: E0302 12:49:43.532041 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.532054 kubelet[3475]: W0302 12:49:43.532050 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.532097 kubelet[3475]: E0302 12:49:43.532061 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.532170 kubelet[3475]: E0302 12:49:43.532157 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.532170 kubelet[3475]: W0302 12:49:43.532166 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.532207 kubelet[3475]: E0302 12:49:43.532172 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.532334 kubelet[3475]: E0302 12:49:43.532320 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.532334 kubelet[3475]: W0302 12:49:43.532329 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.532376 kubelet[3475]: E0302 12:49:43.532336 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.532663 kubelet[3475]: E0302 12:49:43.532555 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.532663 kubelet[3475]: W0302 12:49:43.532566 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.532663 kubelet[3475]: E0302 12:49:43.532577 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.532801 kubelet[3475]: E0302 12:49:43.532789 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.532930 kubelet[3475]: W0302 12:49:43.532836 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.532930 kubelet[3475]: E0302 12:49:43.532850 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.533044 kubelet[3475]: E0302 12:49:43.533033 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.533177 kubelet[3475]: W0302 12:49:43.533084 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.533177 kubelet[3475]: E0302 12:49:43.533097 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.533281 kubelet[3475]: E0302 12:49:43.533271 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.533424 kubelet[3475]: W0302 12:49:43.533319 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.533424 kubelet[3475]: E0302 12:49:43.533332 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.533570 kubelet[3475]: E0302 12:49:43.533559 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.533619 kubelet[3475]: W0302 12:49:43.533610 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.533660 kubelet[3475]: E0302 12:49:43.533649 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.533945 kubelet[3475]: E0302 12:49:43.533926 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.533945 kubelet[3475]: W0302 12:49:43.533939 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.533945 kubelet[3475]: E0302 12:49:43.533948 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.534077 kubelet[3475]: E0302 12:49:43.534064 3475 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 2 12:49:43.534077 kubelet[3475]: W0302 12:49:43.534073 3475 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 2 12:49:43.534123 kubelet[3475]: E0302 12:49:43.534079 3475 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 2 12:49:43.976687 containerd[1899]: time="2026-03-02T12:49:43.976645111Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:43.980298 containerd[1899]: time="2026-03-02T12:49:43.980270867Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3: active requests=0, bytes read=4456989" Mar 2 12:49:43.983344 containerd[1899]: time="2026-03-02T12:49:43.983317370Z" level=info msg="ImageCreate event name:\"sha256:3c477f840adeca332cbee81ef65da50ec7be99ded887a8de75d5cf25b896d6a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:43.987942 containerd[1899]: time="2026-03-02T12:49:43.987913079Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:6cdc6cc2f7cdcbd4bf2d9b6a59c03ed98b5c47f22e467d78b5c06e5fd7bff132\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:43.988680 containerd[1899]: time="2026-03-02T12:49:43.988646680Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\" with image id \"sha256:3c477f840adeca332cbee81ef65da50ec7be99ded887a8de75d5cf25b896d6a9\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:6cdc6cc2f7cdcbd4bf2d9b6a59c03ed98b5c47f22e467d78b5c06e5fd7bff132\", size \"5854474\" in 1.622018093s" Mar 2 12:49:43.988698 containerd[1899]: time="2026-03-02T12:49:43.988683673Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.3\" returns image reference \"sha256:3c477f840adeca332cbee81ef65da50ec7be99ded887a8de75d5cf25b896d6a9\"" Mar 2 12:49:43.996177 containerd[1899]: time="2026-03-02T12:49:43.996145351Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 2 12:49:44.018601 containerd[1899]: time="2026-03-02T12:49:44.018561761Z" level=info msg="Container 0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:44.035934 containerd[1899]: time="2026-03-02T12:49:44.035826413Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2\"" Mar 2 12:49:44.036456 containerd[1899]: time="2026-03-02T12:49:44.036348727Z" level=info msg="StartContainer for \"0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2\"" Mar 2 12:49:44.037625 containerd[1899]: time="2026-03-02T12:49:44.037601825Z" level=info msg="connecting to shim 0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2" address="unix:///run/containerd/s/2b27b6f4cf4c8b74067644f9280c1a5e82cb184a0c3d4e011ad8910a9c617a92" protocol=ttrpc version=3 Mar 2 12:49:44.061618 systemd[1]: Started cri-containerd-0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2.scope - libcontainer container 0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2. Mar 2 12:49:44.135038 containerd[1899]: time="2026-03-02T12:49:44.135001251Z" level=info msg="StartContainer for \"0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2\" returns successfully" Mar 2 12:49:44.139889 systemd[1]: cri-containerd-0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2.scope: Deactivated successfully. Mar 2 12:49:44.143002 containerd[1899]: time="2026-03-02T12:49:44.142975834Z" level=info msg="received container exit event container_id:\"0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2\" id:\"0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2\" pid:4180 exited_at:{seconds:1772455784 nanos:142443576}" Mar 2 12:49:44.158556 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0e606790ff04144159f64b5bf7fa92e6cf7cb7991503afa7ea4efeb4452ea5f2-rootfs.mount: Deactivated successfully. Mar 2 12:49:44.457273 kubelet[3475]: I0302 12:49:44.457247 3475 prober_manager.go:356] "Failed to trigger a manual run" probe="Readiness" Mar 2 12:49:44.476508 kubelet[3475]: I0302 12:49:44.476251 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-typha-744bcc57c6-g97j9" podStartSLOduration=3.010434355 podStartE2EDuration="5.476240988s" podCreationTimestamp="2026-03-02 12:49:39 +0000 UTC" firstStartedPulling="2026-03-02 12:49:39.900414459 +0000 UTC m=+20.602745198" lastFinishedPulling="2026-03-02 12:49:42.366221092 +0000 UTC m=+23.068551831" observedRunningTime="2026-03-02 12:49:43.465030258 +0000 UTC m=+24.167360997" watchObservedRunningTime="2026-03-02 12:49:44.476240988 +0000 UTC m=+25.178571727" Mar 2 12:49:45.380870 kubelet[3475]: E0302 12:49:45.380634 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:45.461830 containerd[1899]: time="2026-03-02T12:49:45.461737491Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.3\"" Mar 2 12:49:47.381523 kubelet[3475]: E0302 12:49:47.381230 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:48.898535 kubelet[3475]: I0302 12:49:48.898254 3475 prober_manager.go:356] "Failed to trigger a manual run" probe="Readiness" Mar 2 12:49:49.381137 kubelet[3475]: E0302 12:49:49.381088 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:50.794428 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1587732648.mount: Deactivated successfully. Mar 2 12:49:51.362310 containerd[1899]: time="2026-03-02T12:49:51.361863416Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:51.365277 containerd[1899]: time="2026-03-02T12:49:51.365256269Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.3: active requests=0, bytes read=153583198" Mar 2 12:49:51.368765 containerd[1899]: time="2026-03-02T12:49:51.368742325Z" level=info msg="ImageCreate event name:\"sha256:98788f64d6cabef718c2551eb8b42ec11d1bfaa912cfeb4f6bf240f79159575d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:51.373080 containerd[1899]: time="2026-03-02T12:49:51.373041272Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:c7aefc80042b94800407ab45640b59402d2897ae8755b9d8370516e7b0e404bc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:51.373501 containerd[1899]: time="2026-03-02T12:49:51.373453469Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.3\" with image id \"sha256:98788f64d6cabef718c2551eb8b42ec11d1bfaa912cfeb4f6bf240f79159575d\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:c7aefc80042b94800407ab45640b59402d2897ae8755b9d8370516e7b0e404bc\", size \"153583060\" in 5.911677473s" Mar 2 12:49:51.373501 containerd[1899]: time="2026-03-02T12:49:51.373479830Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.3\" returns image reference \"sha256:98788f64d6cabef718c2551eb8b42ec11d1bfaa912cfeb4f6bf240f79159575d\"" Mar 2 12:49:51.380353 kubelet[3475]: E0302 12:49:51.380301 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:51.385933 containerd[1899]: time="2026-03-02T12:49:51.385905294Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Mar 2 12:49:51.412359 containerd[1899]: time="2026-03-02T12:49:51.409262390Z" level=info msg="Container cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:51.432503 containerd[1899]: time="2026-03-02T12:49:51.432472546Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477\"" Mar 2 12:49:51.432963 containerd[1899]: time="2026-03-02T12:49:51.432937657Z" level=info msg="StartContainer for \"cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477\"" Mar 2 12:49:51.433939 containerd[1899]: time="2026-03-02T12:49:51.433889055Z" level=info msg="connecting to shim cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477" address="unix:///run/containerd/s/2b27b6f4cf4c8b74067644f9280c1a5e82cb184a0c3d4e011ad8910a9c617a92" protocol=ttrpc version=3 Mar 2 12:49:51.453601 systemd[1]: Started cri-containerd-cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477.scope - libcontainer container cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477. Mar 2 12:49:51.512337 containerd[1899]: time="2026-03-02T12:49:51.512190137Z" level=info msg="StartContainer for \"cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477\" returns successfully" Mar 2 12:49:51.538058 systemd[1]: cri-containerd-cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477.scope: Deactivated successfully. Mar 2 12:49:51.541022 containerd[1899]: time="2026-03-02T12:49:51.540964840Z" level=info msg="received container exit event container_id:\"cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477\" id:\"cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477\" pid:4237 exited_at:{seconds:1772455791 nanos:540778914}" Mar 2 12:49:51.555556 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cdf30014cd8d5a71dbedc2b40cc2e1c7c34ed758f20c27874b489fc145746477-rootfs.mount: Deactivated successfully. Mar 2 12:49:53.381234 kubelet[3475]: E0302 12:49:53.380951 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:53.487893 containerd[1899]: time="2026-03-02T12:49:53.487851813Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.3\"" Mar 2 12:49:55.381027 kubelet[3475]: E0302 12:49:55.380391 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:56.547161 containerd[1899]: time="2026-03-02T12:49:56.546773933Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:56.550234 containerd[1899]: time="2026-03-02T12:49:56.550210314Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.3: active requests=0, bytes read=65998037" Mar 2 12:49:56.553691 containerd[1899]: time="2026-03-02T12:49:56.553663536Z" level=info msg="ImageCreate event name:\"sha256:2aba526dc0b0f95b83ab38a811f41d3daf3ec5ae8876bf273b65b9f142277231\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:56.558028 containerd[1899]: time="2026-03-02T12:49:56.557980643Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:c25deb6a4b79f5e595eb464adf9fb3735ea5623889e249d5b3efa0b42ffcbb47\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:49:56.558416 containerd[1899]: time="2026-03-02T12:49:56.558303870Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.3\" with image id \"sha256:2aba526dc0b0f95b83ab38a811f41d3daf3ec5ae8876bf273b65b9f142277231\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:c25deb6a4b79f5e595eb464adf9fb3735ea5623889e249d5b3efa0b42ffcbb47\", size \"67395562\" in 3.070392136s" Mar 2 12:49:56.558416 containerd[1899]: time="2026-03-02T12:49:56.558328807Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.3\" returns image reference \"sha256:2aba526dc0b0f95b83ab38a811f41d3daf3ec5ae8876bf273b65b9f142277231\"" Mar 2 12:49:56.566652 containerd[1899]: time="2026-03-02T12:49:56.566620090Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 2 12:49:56.593093 containerd[1899]: time="2026-03-02T12:49:56.593065768Z" level=info msg="Container 3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:49:56.595956 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2613995344.mount: Deactivated successfully. Mar 2 12:49:56.611477 containerd[1899]: time="2026-03-02T12:49:56.611447187Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78\"" Mar 2 12:49:56.612022 containerd[1899]: time="2026-03-02T12:49:56.611996838Z" level=info msg="StartContainer for \"3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78\"" Mar 2 12:49:56.612968 containerd[1899]: time="2026-03-02T12:49:56.612939350Z" level=info msg="connecting to shim 3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78" address="unix:///run/containerd/s/2b27b6f4cf4c8b74067644f9280c1a5e82cb184a0c3d4e011ad8910a9c617a92" protocol=ttrpc version=3 Mar 2 12:49:56.629606 systemd[1]: Started cri-containerd-3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78.scope - libcontainer container 3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78. Mar 2 12:49:56.684109 containerd[1899]: time="2026-03-02T12:49:56.684069904Z" level=info msg="StartContainer for \"3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78\" returns successfully" Mar 2 12:49:57.381882 kubelet[3475]: E0302 12:49:57.381846 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:49:59.302304 containerd[1899]: time="2026-03-02T12:49:59.302240390Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 12:49:59.303915 systemd[1]: cri-containerd-3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78.scope: Deactivated successfully. Mar 2 12:49:59.305640 systemd[1]: cri-containerd-3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78.scope: Consumed 336ms CPU time, 190.3M memory peak, 128K read from disk, 171.3M written to disk. Mar 2 12:49:59.306923 containerd[1899]: time="2026-03-02T12:49:59.306765344Z" level=info msg="received container exit event container_id:\"3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78\" id:\"3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78\" pid:4292 exited_at:{seconds:1772455799 nanos:306625003}" Mar 2 12:49:59.321269 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3063595bf2c4f839cf97c7e214c445dc4928e0ee32d4eefe76b47dc809c08d78-rootfs.mount: Deactivated successfully. Mar 2 12:49:59.380860 kubelet[3475]: E0302 12:49:59.380520 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:50:00.382663 kubelet[3475]: I0302 12:49:59.383276 3475 kubelet_node_status.go:427] "Fast updating node status as it just became ready" Mar 2 12:50:00.548071 systemd[1]: Created slice kubepods-besteffort-pod07af8044_862a_4c44_bd2c_12b2e3c44607.slice - libcontainer container kubepods-besteffort-pod07af8044_862a_4c44_bd2c_12b2e3c44607.slice. Mar 2 12:50:00.555089 systemd[1]: Created slice kubepods-burstable-podf9e26872_e431_4fd3_808e_5bafbba0261e.slice - libcontainer container kubepods-burstable-podf9e26872_e431_4fd3_808e_5bafbba0261e.slice. Mar 2 12:50:00.637500 kubelet[3475]: I0302 12:50:00.637246 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9e26872-e431-4fd3-808e-5bafbba0261e-config-volume\") pod \"coredns-7d764666f9-g2d5q\" (UID: \"f9e26872-e431-4fd3-808e-5bafbba0261e\") " pod="kube-system/coredns-7d764666f9-g2d5q" Mar 2 12:50:00.637500 kubelet[3475]: I0302 12:50:00.637275 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x2lp\" (UniqueName: \"kubernetes.io/projected/f9e26872-e431-4fd3-808e-5bafbba0261e-kube-api-access-7x2lp\") pod \"coredns-7d764666f9-g2d5q\" (UID: \"f9e26872-e431-4fd3-808e-5bafbba0261e\") " pod="kube-system/coredns-7d764666f9-g2d5q" Mar 2 12:50:00.645884 systemd[1]: Created slice kubepods-burstable-pod7f11885f_bdcc_4b40_b0d0_d67b83c913b1.slice - libcontainer container kubepods-burstable-pod7f11885f_bdcc_4b40_b0d0_d67b83c913b1.slice. Mar 2 12:50:00.683031 containerd[1899]: time="2026-03-02T12:50:00.682993353Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wtzgs,Uid:07af8044-862a-4c44-bd2c-12b2e3c44607,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:00.737993 kubelet[3475]: I0302 12:50:00.737968 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz56j\" (UniqueName: \"kubernetes.io/projected/7f11885f-bdcc-4b40-b0d0-d67b83c913b1-kube-api-access-gz56j\") pod \"coredns-7d764666f9-rs7zg\" (UID: \"7f11885f-bdcc-4b40-b0d0-d67b83c913b1\") " pod="kube-system/coredns-7d764666f9-rs7zg" Mar 2 12:50:00.738281 kubelet[3475]: I0302 12:50:00.738121 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f11885f-bdcc-4b40-b0d0-d67b83c913b1-config-volume\") pod \"coredns-7d764666f9-rs7zg\" (UID: \"7f11885f-bdcc-4b40-b0d0-d67b83c913b1\") " pod="kube-system/coredns-7d764666f9-rs7zg" Mar 2 12:50:00.738281 kubelet[3475]: I0302 12:50:00.738142 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spjwv\" (UniqueName: \"kubernetes.io/projected/9d4f30bf-d266-4607-b976-3eb09820fbc8-kube-api-access-spjwv\") pod \"calico-apiserver-8487494cdb-p2qk6\" (UID: \"9d4f30bf-d266-4607-b976-3eb09820fbc8\") " pod="calico-system/calico-apiserver-8487494cdb-p2qk6" Mar 2 12:50:00.738281 kubelet[3475]: I0302 12:50:00.738158 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/9d4f30bf-d266-4607-b976-3eb09820fbc8-calico-apiserver-certs\") pod \"calico-apiserver-8487494cdb-p2qk6\" (UID: \"9d4f30bf-d266-4607-b976-3eb09820fbc8\") " pod="calico-system/calico-apiserver-8487494cdb-p2qk6" Mar 2 12:50:00.740653 systemd[1]: Created slice kubepods-besteffort-pod9d4f30bf_d266_4607_b976_3eb09820fbc8.slice - libcontainer container kubepods-besteffort-pod9d4f30bf_d266_4607_b976_3eb09820fbc8.slice. Mar 2 12:50:00.813053 systemd[1]: Created slice kubepods-besteffort-podb7cc8850_f440_4499_b33a_69338a3f54c1.slice - libcontainer container kubepods-besteffort-podb7cc8850_f440_4499_b33a_69338a3f54c1.slice. Mar 2 12:50:00.820429 systemd[1]: Created slice kubepods-besteffort-poda6e52250_759f_484f_86cd_140ce9b02f00.slice - libcontainer container kubepods-besteffort-poda6e52250_759f_484f_86cd_140ce9b02f00.slice. Mar 2 12:50:00.840027 kubelet[3475]: I0302 12:50:00.839291 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7ndb\" (UniqueName: \"kubernetes.io/projected/a6e52250-759f-484f-86cd-140ce9b02f00-kube-api-access-g7ndb\") pod \"calico-apiserver-68bfcfb4bc-jjtz2\" (UID: \"a6e52250-759f-484f-86cd-140ce9b02f00\") " pod="calico-system/calico-apiserver-68bfcfb4bc-jjtz2" Mar 2 12:50:00.840027 kubelet[3475]: I0302 12:50:00.839331 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7cc8850-f440-4499-b33a-69338a3f54c1-tigera-ca-bundle\") pod \"calico-kube-controllers-6b7cf495fc-ftpl9\" (UID: \"b7cc8850-f440-4499-b33a-69338a3f54c1\") " pod="calico-system/calico-kube-controllers-6b7cf495fc-ftpl9" Mar 2 12:50:00.840027 kubelet[3475]: I0302 12:50:00.839351 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/a6e52250-759f-484f-86cd-140ce9b02f00-calico-apiserver-certs\") pod \"calico-apiserver-68bfcfb4bc-jjtz2\" (UID: \"a6e52250-759f-484f-86cd-140ce9b02f00\") " pod="calico-system/calico-apiserver-68bfcfb4bc-jjtz2" Mar 2 12:50:00.840027 kubelet[3475]: I0302 12:50:00.839373 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh6z7\" (UniqueName: \"kubernetes.io/projected/b7cc8850-f440-4499-b33a-69338a3f54c1-kube-api-access-kh6z7\") pod \"calico-kube-controllers-6b7cf495fc-ftpl9\" (UID: \"b7cc8850-f440-4499-b33a-69338a3f54c1\") " pod="calico-system/calico-kube-controllers-6b7cf495fc-ftpl9" Mar 2 12:50:00.865315 systemd[1]: Created slice kubepods-besteffort-podcfb890bf_40cf_4b4a_b482_8dad57b17735.slice - libcontainer container kubepods-besteffort-podcfb890bf_40cf_4b4a_b482_8dad57b17735.slice. Mar 2 12:50:00.876239 containerd[1899]: time="2026-03-02T12:50:00.876195320Z" level=error msg="Failed to destroy network for sandbox \"3a2443f0abbbdffcd5b402be8d1b39b04ee4ce4ff2b51fb67f5374003950fdc1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:00.879792 systemd[1]: Created slice kubepods-besteffort-pod56b6d31a_681f_4836_a653_77f50d324a86.slice - libcontainer container kubepods-besteffort-pod56b6d31a_681f_4836_a653_77f50d324a86.slice. Mar 2 12:50:00.935383 containerd[1899]: time="2026-03-02T12:50:00.935281879Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wtzgs,Uid:07af8044-862a-4c44-bd2c-12b2e3c44607,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3a2443f0abbbdffcd5b402be8d1b39b04ee4ce4ff2b51fb67f5374003950fdc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:00.936664 kubelet[3475]: E0302 12:50:00.936626 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3a2443f0abbbdffcd5b402be8d1b39b04ee4ce4ff2b51fb67f5374003950fdc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:00.936736 kubelet[3475]: E0302 12:50:00.936673 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3a2443f0abbbdffcd5b402be8d1b39b04ee4ce4ff2b51fb67f5374003950fdc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-wtzgs" Mar 2 12:50:00.936736 kubelet[3475]: E0302 12:50:00.936686 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3a2443f0abbbdffcd5b402be8d1b39b04ee4ce4ff2b51fb67f5374003950fdc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-wtzgs" Mar 2 12:50:00.936736 kubelet[3475]: E0302 12:50:00.936717 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-wtzgs_calico-system(07af8044-862a-4c44-bd2c-12b2e3c44607)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-wtzgs_calico-system(07af8044-862a-4c44-bd2c-12b2e3c44607)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3a2443f0abbbdffcd5b402be8d1b39b04ee4ce4ff2b51fb67f5374003950fdc1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-wtzgs" podUID="07af8044-862a-4c44-bd2c-12b2e3c44607" Mar 2 12:50:00.939631 containerd[1899]: time="2026-03-02T12:50:00.939525448Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-g2d5q,Uid:f9e26872-e431-4fd3-808e-5bafbba0261e,Namespace:kube-system,Attempt:0,}" Mar 2 12:50:00.940258 kubelet[3475]: I0302 12:50:00.940005 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb890bf-40cf-4b4a-b482-8dad57b17735-goldmane-ca-bundle\") pod \"goldmane-7d7658d587-6g28g\" (UID: \"cfb890bf-40cf-4b4a-b482-8dad57b17735\") " pod="calico-system/goldmane-7d7658d587-6g28g" Mar 2 12:50:00.940258 kubelet[3475]: I0302 12:50:00.940038 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/cfb890bf-40cf-4b4a-b482-8dad57b17735-goldmane-key-pair\") pod \"goldmane-7d7658d587-6g28g\" (UID: \"cfb890bf-40cf-4b4a-b482-8dad57b17735\") " pod="calico-system/goldmane-7d7658d587-6g28g" Mar 2 12:50:00.940258 kubelet[3475]: I0302 12:50:00.940191 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfb890bf-40cf-4b4a-b482-8dad57b17735-config\") pod \"goldmane-7d7658d587-6g28g\" (UID: \"cfb890bf-40cf-4b4a-b482-8dad57b17735\") " pod="calico-system/goldmane-7d7658d587-6g28g" Mar 2 12:50:00.941283 kubelet[3475]: I0302 12:50:00.940309 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96cz\" (UniqueName: \"kubernetes.io/projected/cfb890bf-40cf-4b4a-b482-8dad57b17735-kube-api-access-h96cz\") pod \"goldmane-7d7658d587-6g28g\" (UID: \"cfb890bf-40cf-4b4a-b482-8dad57b17735\") " pod="calico-system/goldmane-7d7658d587-6g28g" Mar 2 12:50:00.941283 kubelet[3475]: I0302 12:50:00.940398 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/56b6d31a-681f-4836-a653-77f50d324a86-calico-apiserver-certs\") pod \"calico-apiserver-8487494cdb-7zh2j\" (UID: \"56b6d31a-681f-4836-a653-77f50d324a86\") " pod="calico-system/calico-apiserver-8487494cdb-7zh2j" Mar 2 12:50:00.941283 kubelet[3475]: I0302 12:50:00.940418 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzv8c\" (UniqueName: \"kubernetes.io/projected/56b6d31a-681f-4836-a653-77f50d324a86-kube-api-access-dzv8c\") pod \"calico-apiserver-8487494cdb-7zh2j\" (UID: \"56b6d31a-681f-4836-a653-77f50d324a86\") " pod="calico-system/calico-apiserver-8487494cdb-7zh2j" Mar 2 12:50:00.943468 systemd[1]: Created slice kubepods-besteffort-pod754476fb_8279_47ac_97a1_35afa4b3c0b6.slice - libcontainer container kubepods-besteffort-pod754476fb_8279_47ac_97a1_35afa4b3c0b6.slice. Mar 2 12:50:00.987255 containerd[1899]: time="2026-03-02T12:50:00.987231755Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rs7zg,Uid:7f11885f-bdcc-4b40-b0d0-d67b83c913b1,Namespace:kube-system,Attempt:0,}" Mar 2 12:50:01.040907 kubelet[3475]: I0302 12:50:01.040852 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzn75\" (UniqueName: \"kubernetes.io/projected/754476fb-8279-47ac-97a1-35afa4b3c0b6-kube-api-access-nzn75\") pod \"whisker-cdd6698c4-hjntz\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " pod="calico-system/whisker-cdd6698c4-hjntz" Mar 2 12:50:01.041526 kubelet[3475]: I0302 12:50:01.041028 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-nginx-config\") pod \"whisker-cdd6698c4-hjntz\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " pod="calico-system/whisker-cdd6698c4-hjntz" Mar 2 12:50:01.041526 kubelet[3475]: I0302 12:50:01.041050 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-backend-key-pair\") pod \"whisker-cdd6698c4-hjntz\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " pod="calico-system/whisker-cdd6698c4-hjntz" Mar 2 12:50:01.041526 kubelet[3475]: I0302 12:50:01.041070 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-ca-bundle\") pod \"whisker-cdd6698c4-hjntz\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " pod="calico-system/whisker-cdd6698c4-hjntz" Mar 2 12:50:01.115944 containerd[1899]: time="2026-03-02T12:50:01.115886936Z" level=error msg="Failed to destroy network for sandbox \"6fe8d1e53ff58791a5c50677b8a5f7080f5c8b87c14984697136dc40eab7ff3c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.190821 containerd[1899]: time="2026-03-02T12:50:01.190342147Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-p2qk6,Uid:9d4f30bf-d266-4607-b976-3eb09820fbc8,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:01.220453 containerd[1899]: time="2026-03-02T12:50:01.220339987Z" level=error msg="Failed to destroy network for sandbox \"f1643b43cb8ef90850f189ef5f39fd827750e4ed3ffccd3e0de149e838552275\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.236382 containerd[1899]: time="2026-03-02T12:50:01.236326740Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-g2d5q,Uid:f9e26872-e431-4fd3-808e-5bafbba0261e,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"6fe8d1e53ff58791a5c50677b8a5f7080f5c8b87c14984697136dc40eab7ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.237393 containerd[1899]: time="2026-03-02T12:50:01.237219243Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6b7cf495fc-ftpl9,Uid:b7cc8850-f440-4499-b33a-69338a3f54c1,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:01.238346 kubelet[3475]: E0302 12:50:01.237555 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6fe8d1e53ff58791a5c50677b8a5f7080f5c8b87c14984697136dc40eab7ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.238346 kubelet[3475]: E0302 12:50:01.237599 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6fe8d1e53ff58791a5c50677b8a5f7080f5c8b87c14984697136dc40eab7ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-g2d5q" Mar 2 12:50:01.238346 kubelet[3475]: E0302 12:50:01.237611 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6fe8d1e53ff58791a5c50677b8a5f7080f5c8b87c14984697136dc40eab7ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-g2d5q" Mar 2 12:50:01.238484 kubelet[3475]: E0302 12:50:01.237647 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7d764666f9-g2d5q_kube-system(f9e26872-e431-4fd3-808e-5bafbba0261e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7d764666f9-g2d5q_kube-system(f9e26872-e431-4fd3-808e-5bafbba0261e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6fe8d1e53ff58791a5c50677b8a5f7080f5c8b87c14984697136dc40eab7ff3c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-g2d5q" podUID="f9e26872-e431-4fd3-808e-5bafbba0261e" Mar 2 12:50:01.286650 containerd[1899]: time="2026-03-02T12:50:01.286632488Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-7zh2j,Uid:56b6d31a-681f-4836-a653-77f50d324a86,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:01.331823 containerd[1899]: time="2026-03-02T12:50:01.331793630Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rs7zg,Uid:7f11885f-bdcc-4b40-b0d0-d67b83c913b1,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1643b43cb8ef90850f189ef5f39fd827750e4ed3ffccd3e0de149e838552275\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.332176 kubelet[3475]: E0302 12:50:01.332146 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1643b43cb8ef90850f189ef5f39fd827750e4ed3ffccd3e0de149e838552275\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.332238 kubelet[3475]: E0302 12:50:01.332187 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1643b43cb8ef90850f189ef5f39fd827750e4ed3ffccd3e0de149e838552275\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-rs7zg" Mar 2 12:50:01.332238 kubelet[3475]: E0302 12:50:01.332200 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1643b43cb8ef90850f189ef5f39fd827750e4ed3ffccd3e0de149e838552275\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7d764666f9-rs7zg" Mar 2 12:50:01.332275 kubelet[3475]: E0302 12:50:01.332232 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7d764666f9-rs7zg_kube-system(7f11885f-bdcc-4b40-b0d0-d67b83c913b1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7d764666f9-rs7zg_kube-system(7f11885f-bdcc-4b40-b0d0-d67b83c913b1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f1643b43cb8ef90850f189ef5f39fd827750e4ed3ffccd3e0de149e838552275\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7d764666f9-rs7zg" podUID="7f11885f-bdcc-4b40-b0d0-d67b83c913b1" Mar 2 12:50:01.332464 containerd[1899]: time="2026-03-02T12:50:01.332443308Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68bfcfb4bc-jjtz2,Uid:a6e52250-759f-484f-86cd-140ce9b02f00,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:01.383084 containerd[1899]: time="2026-03-02T12:50:01.383061762Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7d7658d587-6g28g,Uid:cfb890bf-40cf-4b4a-b482-8dad57b17735,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:01.449253 containerd[1899]: time="2026-03-02T12:50:01.448575876Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-cdd6698c4-hjntz,Uid:754476fb-8279-47ac-97a1-35afa4b3c0b6,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:01.478007 containerd[1899]: time="2026-03-02T12:50:01.477974727Z" level=error msg="Failed to destroy network for sandbox \"e8ab2f6768a3b5ae511e586981a1fc378bc19cea2b10991ba10b477f092f6145\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.544352 containerd[1899]: time="2026-03-02T12:50:01.543909064Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 2 12:50:01.668784 containerd[1899]: time="2026-03-02T12:50:01.668725130Z" level=error msg="Failed to destroy network for sandbox \"60be8a2e0d0b40a6c72abe2310523fbeaf9ee8280f942b700326c200798642e3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.729575 containerd[1899]: time="2026-03-02T12:50:01.728992378Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-p2qk6,Uid:9d4f30bf-d266-4607-b976-3eb09820fbc8,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8ab2f6768a3b5ae511e586981a1fc378bc19cea2b10991ba10b477f092f6145\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.729865 kubelet[3475]: E0302 12:50:01.729206 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8ab2f6768a3b5ae511e586981a1fc378bc19cea2b10991ba10b477f092f6145\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.729865 kubelet[3475]: E0302 12:50:01.729284 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8ab2f6768a3b5ae511e586981a1fc378bc19cea2b10991ba10b477f092f6145\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-8487494cdb-p2qk6" Mar 2 12:50:01.729865 kubelet[3475]: E0302 12:50:01.729298 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8ab2f6768a3b5ae511e586981a1fc378bc19cea2b10991ba10b477f092f6145\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-8487494cdb-p2qk6" Mar 2 12:50:01.730036 kubelet[3475]: E0302 12:50:01.729351 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-8487494cdb-p2qk6_calico-system(9d4f30bf-d266-4607-b976-3eb09820fbc8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-8487494cdb-p2qk6_calico-system(9d4f30bf-d266-4607-b976-3eb09820fbc8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e8ab2f6768a3b5ae511e586981a1fc378bc19cea2b10991ba10b477f092f6145\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-8487494cdb-p2qk6" podUID="9d4f30bf-d266-4607-b976-3eb09820fbc8" Mar 2 12:50:01.741604 systemd[1]: run-netns-cni\x2de588b7e7\x2dff12\x2dd6ee\x2d7fc3\x2d8319bddf3999.mount: Deactivated successfully. Mar 2 12:50:01.770848 containerd[1899]: time="2026-03-02T12:50:01.770809260Z" level=error msg="Failed to destroy network for sandbox \"38bb5ef0e23074e60ec0576d081e4e7fa95088665537219e40d43f321190e072\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.772209 systemd[1]: run-netns-cni\x2da70217bf\x2db0eb\x2d4870\x2d466b\x2da5825146f816.mount: Deactivated successfully. Mar 2 12:50:01.824915 containerd[1899]: time="2026-03-02T12:50:01.824884608Z" level=error msg="Failed to destroy network for sandbox \"e786fde79f6bc5d208d41e5156e482962b5a1ff11e8599402844ffc6a73a03a0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.826225 systemd[1]: run-netns-cni\x2d0dd0d151\x2dd4ee\x2ddf2f\x2d9c30\x2ddda67e5589d3.mount: Deactivated successfully. Mar 2 12:50:01.876568 containerd[1899]: time="2026-03-02T12:50:01.875209373Z" level=error msg="Failed to destroy network for sandbox \"19b6239b3471f6e8f32ff3efa1ad347114516b4f2d0297c2e36bcf8db6c61ca3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.876563 systemd[1]: run-netns-cni\x2de49d3b97\x2d558c\x2d0743\x2d3af5\x2dd3a90e01f6e0.mount: Deactivated successfully. Mar 2 12:50:01.886975 containerd[1899]: time="2026-03-02T12:50:01.886885331Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6b7cf495fc-ftpl9,Uid:b7cc8850-f440-4499-b33a-69338a3f54c1,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"60be8a2e0d0b40a6c72abe2310523fbeaf9ee8280f942b700326c200798642e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.887120 kubelet[3475]: E0302 12:50:01.887080 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"60be8a2e0d0b40a6c72abe2310523fbeaf9ee8280f942b700326c200798642e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.887185 kubelet[3475]: E0302 12:50:01.887128 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"60be8a2e0d0b40a6c72abe2310523fbeaf9ee8280f942b700326c200798642e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6b7cf495fc-ftpl9" Mar 2 12:50:01.887185 kubelet[3475]: E0302 12:50:01.887142 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"60be8a2e0d0b40a6c72abe2310523fbeaf9ee8280f942b700326c200798642e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6b7cf495fc-ftpl9" Mar 2 12:50:01.887231 kubelet[3475]: E0302 12:50:01.887179 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6b7cf495fc-ftpl9_calico-system(b7cc8850-f440-4499-b33a-69338a3f54c1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6b7cf495fc-ftpl9_calico-system(b7cc8850-f440-4499-b33a-69338a3f54c1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"60be8a2e0d0b40a6c72abe2310523fbeaf9ee8280f942b700326c200798642e3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6b7cf495fc-ftpl9" podUID="b7cc8850-f440-4499-b33a-69338a3f54c1" Mar 2 12:50:01.921560 containerd[1899]: time="2026-03-02T12:50:01.921527529Z" level=error msg="Failed to destroy network for sandbox \"dbe1680051b22fcf070d8bbbac43d132b8c82d64f882f12076f38b2b40944723\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.923103 systemd[1]: run-netns-cni\x2d59a14f1a\x2de0dd\x2d540c\x2d9fa2\x2da646f926c412.mount: Deactivated successfully. Mar 2 12:50:01.936082 containerd[1899]: time="2026-03-02T12:50:01.936028856Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-cdd6698c4-hjntz,Uid:754476fb-8279-47ac-97a1-35afa4b3c0b6,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"38bb5ef0e23074e60ec0576d081e4e7fa95088665537219e40d43f321190e072\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.936622 kubelet[3475]: E0302 12:50:01.936315 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38bb5ef0e23074e60ec0576d081e4e7fa95088665537219e40d43f321190e072\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:01.936622 kubelet[3475]: E0302 12:50:01.936362 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38bb5ef0e23074e60ec0576d081e4e7fa95088665537219e40d43f321190e072\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-cdd6698c4-hjntz" Mar 2 12:50:01.936622 kubelet[3475]: E0302 12:50:01.936388 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38bb5ef0e23074e60ec0576d081e4e7fa95088665537219e40d43f321190e072\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-cdd6698c4-hjntz" Mar 2 12:50:01.936699 kubelet[3475]: E0302 12:50:01.936428 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-cdd6698c4-hjntz_calico-system(754476fb-8279-47ac-97a1-35afa4b3c0b6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-cdd6698c4-hjntz_calico-system(754476fb-8279-47ac-97a1-35afa4b3c0b6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"38bb5ef0e23074e60ec0576d081e4e7fa95088665537219e40d43f321190e072\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-cdd6698c4-hjntz" podUID="754476fb-8279-47ac-97a1-35afa4b3c0b6" Mar 2 12:50:02.035340 containerd[1899]: time="2026-03-02T12:50:02.035184254Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-7zh2j,Uid:56b6d31a-681f-4836-a653-77f50d324a86,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e786fde79f6bc5d208d41e5156e482962b5a1ff11e8599402844ffc6a73a03a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:02.035725 kubelet[3475]: E0302 12:50:02.035686 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e786fde79f6bc5d208d41e5156e482962b5a1ff11e8599402844ffc6a73a03a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:02.035924 kubelet[3475]: E0302 12:50:02.035864 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e786fde79f6bc5d208d41e5156e482962b5a1ff11e8599402844ffc6a73a03a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-8487494cdb-7zh2j" Mar 2 12:50:02.036056 kubelet[3475]: E0302 12:50:02.035908 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e786fde79f6bc5d208d41e5156e482962b5a1ff11e8599402844ffc6a73a03a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-8487494cdb-7zh2j" Mar 2 12:50:02.036136 kubelet[3475]: E0302 12:50:02.036037 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-8487494cdb-7zh2j_calico-system(56b6d31a-681f-4836-a653-77f50d324a86)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-8487494cdb-7zh2j_calico-system(56b6d31a-681f-4836-a653-77f50d324a86)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e786fde79f6bc5d208d41e5156e482962b5a1ff11e8599402844ffc6a73a03a0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-8487494cdb-7zh2j" podUID="56b6d31a-681f-4836-a653-77f50d324a86" Mar 2 12:50:02.096105 containerd[1899]: time="2026-03-02T12:50:02.096028929Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68bfcfb4bc-jjtz2,Uid:a6e52250-759f-484f-86cd-140ce9b02f00,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"19b6239b3471f6e8f32ff3efa1ad347114516b4f2d0297c2e36bcf8db6c61ca3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:02.096199 kubelet[3475]: E0302 12:50:02.096170 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"19b6239b3471f6e8f32ff3efa1ad347114516b4f2d0297c2e36bcf8db6c61ca3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:02.096226 kubelet[3475]: E0302 12:50:02.096199 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"19b6239b3471f6e8f32ff3efa1ad347114516b4f2d0297c2e36bcf8db6c61ca3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-68bfcfb4bc-jjtz2" Mar 2 12:50:02.096226 kubelet[3475]: E0302 12:50:02.096210 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"19b6239b3471f6e8f32ff3efa1ad347114516b4f2d0297c2e36bcf8db6c61ca3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-68bfcfb4bc-jjtz2" Mar 2 12:50:02.096318 kubelet[3475]: E0302 12:50:02.096252 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-68bfcfb4bc-jjtz2_calico-system(a6e52250-759f-484f-86cd-140ce9b02f00)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-68bfcfb4bc-jjtz2_calico-system(a6e52250-759f-484f-86cd-140ce9b02f00)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"19b6239b3471f6e8f32ff3efa1ad347114516b4f2d0297c2e36bcf8db6c61ca3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-68bfcfb4bc-jjtz2" podUID="a6e52250-759f-484f-86cd-140ce9b02f00" Mar 2 12:50:02.103012 containerd[1899]: time="2026-03-02T12:50:02.102929101Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7d7658d587-6g28g,Uid:cfb890bf-40cf-4b4a-b482-8dad57b17735,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"dbe1680051b22fcf070d8bbbac43d132b8c82d64f882f12076f38b2b40944723\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:02.103093 kubelet[3475]: E0302 12:50:02.103051 3475 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dbe1680051b22fcf070d8bbbac43d132b8c82d64f882f12076f38b2b40944723\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 2 12:50:02.103093 kubelet[3475]: E0302 12:50:02.103077 3475 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dbe1680051b22fcf070d8bbbac43d132b8c82d64f882f12076f38b2b40944723\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7d7658d587-6g28g" Mar 2 12:50:02.103093 kubelet[3475]: E0302 12:50:02.103087 3475 kuberuntime_manager.go:1558] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dbe1680051b22fcf070d8bbbac43d132b8c82d64f882f12076f38b2b40944723\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7d7658d587-6g28g" Mar 2 12:50:02.103156 kubelet[3475]: E0302 12:50:02.103112 3475 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7d7658d587-6g28g_calico-system(cfb890bf-40cf-4b4a-b482-8dad57b17735)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7d7658d587-6g28g_calico-system(cfb890bf-40cf-4b4a-b482-8dad57b17735)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dbe1680051b22fcf070d8bbbac43d132b8c82d64f882f12076f38b2b40944723\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7d7658d587-6g28g" podUID="cfb890bf-40cf-4b4a-b482-8dad57b17735" Mar 2 12:50:02.142531 containerd[1899]: time="2026-03-02T12:50:02.141959472Z" level=info msg="Container 9c2832a961e6508cf5c06ce7404d4b7f5f26f32fd2173a1a81205c343698a7ee: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:02.290396 containerd[1899]: time="2026-03-02T12:50:02.290314141Z" level=info msg="CreateContainer within sandbox \"5ebe2d057571c757ff13418cd96132315902ca24046016f4b1fbf33dbdcd7539\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"9c2832a961e6508cf5c06ce7404d4b7f5f26f32fd2173a1a81205c343698a7ee\"" Mar 2 12:50:02.291798 containerd[1899]: time="2026-03-02T12:50:02.291773046Z" level=info msg="StartContainer for \"9c2832a961e6508cf5c06ce7404d4b7f5f26f32fd2173a1a81205c343698a7ee\"" Mar 2 12:50:02.293455 containerd[1899]: time="2026-03-02T12:50:02.293421223Z" level=info msg="connecting to shim 9c2832a961e6508cf5c06ce7404d4b7f5f26f32fd2173a1a81205c343698a7ee" address="unix:///run/containerd/s/2b27b6f4cf4c8b74067644f9280c1a5e82cb184a0c3d4e011ad8910a9c617a92" protocol=ttrpc version=3 Mar 2 12:50:02.310608 systemd[1]: Started cri-containerd-9c2832a961e6508cf5c06ce7404d4b7f5f26f32fd2173a1a81205c343698a7ee.scope - libcontainer container 9c2832a961e6508cf5c06ce7404d4b7f5f26f32fd2173a1a81205c343698a7ee. Mar 2 12:50:02.373519 containerd[1899]: time="2026-03-02T12:50:02.372817843Z" level=info msg="StartContainer for \"9c2832a961e6508cf5c06ce7404d4b7f5f26f32fd2173a1a81205c343698a7ee\" returns successfully" Mar 2 12:50:02.552011 kubelet[3475]: I0302 12:50:02.551914 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/754476fb-8279-47ac-97a1-35afa4b3c0b6-kube-api-access-nzn75\" (UniqueName: \"kubernetes.io/projected/754476fb-8279-47ac-97a1-35afa4b3c0b6-kube-api-access-nzn75\") pod \"754476fb-8279-47ac-97a1-35afa4b3c0b6\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " Mar 2 12:50:02.552011 kubelet[3475]: I0302 12:50:02.551954 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-nginx-config\" (UniqueName: \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-nginx-config\") pod \"754476fb-8279-47ac-97a1-35afa4b3c0b6\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " Mar 2 12:50:02.552011 kubelet[3475]: I0302 12:50:02.551973 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-ca-bundle\") pod \"754476fb-8279-47ac-97a1-35afa4b3c0b6\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " Mar 2 12:50:02.552011 kubelet[3475]: I0302 12:50:02.551990 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/secret/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-backend-key-pair\") pod \"754476fb-8279-47ac-97a1-35afa4b3c0b6\" (UID: \"754476fb-8279-47ac-97a1-35afa4b3c0b6\") " Mar 2 12:50:02.552855 kubelet[3475]: I0302 12:50:02.552812 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-nginx-config" pod "754476fb-8279-47ac-97a1-35afa4b3c0b6" (UID: "754476fb-8279-47ac-97a1-35afa4b3c0b6"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 12:50:02.553155 kubelet[3475]: I0302 12:50:02.553094 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-ca-bundle" pod "754476fb-8279-47ac-97a1-35afa4b3c0b6" (UID: "754476fb-8279-47ac-97a1-35afa4b3c0b6"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 12:50:02.557772 kubelet[3475]: I0302 12:50:02.557282 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-node-h5k9w" podStartSLOduration=2.044345104 podStartE2EDuration="23.557273463s" podCreationTimestamp="2026-03-02 12:49:39 +0000 UTC" firstStartedPulling="2026-03-02 12:49:39.993187128 +0000 UTC m=+20.695517875" lastFinishedPulling="2026-03-02 12:50:01.506115495 +0000 UTC m=+42.208446234" observedRunningTime="2026-03-02 12:50:02.537986165 +0000 UTC m=+43.240316904" watchObservedRunningTime="2026-03-02 12:50:02.557273463 +0000 UTC m=+43.259604210" Mar 2 12:50:02.560071 kubelet[3475]: I0302 12:50:02.559985 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/754476fb-8279-47ac-97a1-35afa4b3c0b6-kube-api-access-nzn75" pod "754476fb-8279-47ac-97a1-35afa4b3c0b6" (UID: "754476fb-8279-47ac-97a1-35afa4b3c0b6"). InnerVolumeSpecName "kube-api-access-nzn75". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 12:50:02.561338 kubelet[3475]: I0302 12:50:02.561304 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-backend-key-pair" pod "754476fb-8279-47ac-97a1-35afa4b3c0b6" (UID: "754476fb-8279-47ac-97a1-35afa4b3c0b6"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 2 12:50:02.653323 kubelet[3475]: I0302 12:50:02.653290 3475 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nzn75\" (UniqueName: \"kubernetes.io/projected/754476fb-8279-47ac-97a1-35afa4b3c0b6-kube-api-access-nzn75\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:02.653323 kubelet[3475]: I0302 12:50:02.653320 3475 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-nginx-config\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:02.653323 kubelet[3475]: I0302 12:50:02.653326 3475 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-ca-bundle\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:02.653323 kubelet[3475]: I0302 12:50:02.653335 3475 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/754476fb-8279-47ac-97a1-35afa4b3c0b6-whisker-backend-key-pair\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:02.734707 systemd[1]: var-lib-kubelet-pods-754476fb\x2d8279\x2d47ac\x2d97a1\x2d35afa4b3c0b6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dnzn75.mount: Deactivated successfully. Mar 2 12:50:02.734781 systemd[1]: var-lib-kubelet-pods-754476fb\x2d8279\x2d47ac\x2d97a1\x2d35afa4b3c0b6-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Mar 2 12:50:03.386035 systemd[1]: Removed slice kubepods-besteffort-pod754476fb_8279_47ac_97a1_35afa4b3c0b6.slice - libcontainer container kubepods-besteffort-pod754476fb_8279_47ac_97a1_35afa4b3c0b6.slice. Mar 2 12:50:03.600536 systemd[1]: Created slice kubepods-besteffort-pod3dfbabb8_d88d_426e_b3c4_32a46e087502.slice - libcontainer container kubepods-besteffort-pod3dfbabb8_d88d_426e_b3c4_32a46e087502.slice. Mar 2 12:50:03.660577 kubelet[3475]: I0302 12:50:03.660462 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhjb9\" (UniqueName: \"kubernetes.io/projected/3dfbabb8-d88d-426e-b3c4-32a46e087502-kube-api-access-jhjb9\") pod \"whisker-cd5749c56-698wr\" (UID: \"3dfbabb8-d88d-426e-b3c4-32a46e087502\") " pod="calico-system/whisker-cd5749c56-698wr" Mar 2 12:50:03.661048 kubelet[3475]: I0302 12:50:03.660964 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/3dfbabb8-d88d-426e-b3c4-32a46e087502-whisker-backend-key-pair\") pod \"whisker-cd5749c56-698wr\" (UID: \"3dfbabb8-d88d-426e-b3c4-32a46e087502\") " pod="calico-system/whisker-cd5749c56-698wr" Mar 2 12:50:03.661048 kubelet[3475]: I0302 12:50:03.660989 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/3dfbabb8-d88d-426e-b3c4-32a46e087502-nginx-config\") pod \"whisker-cd5749c56-698wr\" (UID: \"3dfbabb8-d88d-426e-b3c4-32a46e087502\") " pod="calico-system/whisker-cd5749c56-698wr" Mar 2 12:50:03.661048 kubelet[3475]: I0302 12:50:03.661000 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3dfbabb8-d88d-426e-b3c4-32a46e087502-whisker-ca-bundle\") pod \"whisker-cd5749c56-698wr\" (UID: \"3dfbabb8-d88d-426e-b3c4-32a46e087502\") " pod="calico-system/whisker-cd5749c56-698wr" Mar 2 12:50:04.993675 containerd[1899]: time="2026-03-02T12:50:04.993632514Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-cd5749c56-698wr,Uid:3dfbabb8-d88d-426e-b3c4-32a46e087502,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:05.383861 kubelet[3475]: I0302 12:50:05.383827 3475 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="754476fb-8279-47ac-97a1-35afa4b3c0b6" path="/var/lib/kubelet/pods/754476fb-8279-47ac-97a1-35afa4b3c0b6/volumes" Mar 2 12:50:05.807366 systemd-networkd[1485]: vxlan.calico: Link UP Mar 2 12:50:05.807375 systemd-networkd[1485]: vxlan.calico: Gained carrier Mar 2 12:50:06.152673 systemd-networkd[1485]: cali44b9a32bd69: Link UP Mar 2 12:50:06.153370 systemd-networkd[1485]: cali44b9a32bd69: Gained carrier Mar 2 12:50:06.183828 containerd[1899]: 2026-03-02 12:50:06.088 [INFO][4833] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0 whisker-cd5749c56- calico-system 3dfbabb8-d88d-426e-b3c4-32a46e087502 938 0 2026-03-02 12:50:03 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:cd5749c56 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 whisker-cd5749c56-698wr eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali44b9a32bd69 [] [] }} ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-" Mar 2 12:50:06.183828 containerd[1899]: 2026-03-02 12:50:06.090 [INFO][4833] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" Mar 2 12:50:06.183828 containerd[1899]: 2026-03-02 12:50:06.107 [INFO][4845] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" HandleID="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Workload="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.112 [INFO][4845] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" HandleID="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Workload="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ed4b0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"whisker-cd5749c56-698wr", "timestamp":"2026-03-02 12:50:06.107390597 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000333080)} Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.112 [INFO][4845] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.112 [INFO][4845] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.112 [INFO][4845] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.113 [INFO][4845] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.117 [INFO][4845] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.121 [INFO][4845] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.122 [INFO][4845] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248631 containerd[1899]: 2026-03-02 12:50:06.123 [INFO][4845] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248800 containerd[1899]: 2026-03-02 12:50:06.123 [INFO][4845] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248800 containerd[1899]: 2026-03-02 12:50:06.124 [INFO][4845] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a Mar 2 12:50:06.248800 containerd[1899]: 2026-03-02 12:50:06.131 [INFO][4845] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248800 containerd[1899]: 2026-03-02 12:50:06.136 [INFO][4845] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.65/26] block=192.168.54.64/26 handle="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248800 containerd[1899]: 2026-03-02 12:50:06.136 [INFO][4845] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.65/26] handle="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" host="ci-4459.2.101-47783670b7" Mar 2 12:50:06.248800 containerd[1899]: 2026-03-02 12:50:06.136 [INFO][4845] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:06.248800 containerd[1899]: 2026-03-02 12:50:06.136 [INFO][4845] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.65/26] IPv6=[] ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" HandleID="k8s-pod-network.52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Workload="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" Mar 2 12:50:06.248894 containerd[1899]: 2026-03-02 12:50:06.139 [INFO][4833] cni-plugin/k8s.go 418: Populated endpoint ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0", GenerateName:"whisker-cd5749c56-", Namespace:"calico-system", SelfLink:"", UID:"3dfbabb8-d88d-426e-b3c4-32a46e087502", ResourceVersion:"938", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 50, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"cd5749c56", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"whisker-cd5749c56-698wr", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.54.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali44b9a32bd69", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:06.248894 containerd[1899]: 2026-03-02 12:50:06.139 [INFO][4833] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.65/32] ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" Mar 2 12:50:06.248943 containerd[1899]: 2026-03-02 12:50:06.139 [INFO][4833] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali44b9a32bd69 ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" Mar 2 12:50:06.248943 containerd[1899]: 2026-03-02 12:50:06.154 [INFO][4833] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" Mar 2 12:50:06.248972 containerd[1899]: 2026-03-02 12:50:06.154 [INFO][4833] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0", GenerateName:"whisker-cd5749c56-", Namespace:"calico-system", SelfLink:"", UID:"3dfbabb8-d88d-426e-b3c4-32a46e087502", ResourceVersion:"938", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 50, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"cd5749c56", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a", Pod:"whisker-cd5749c56-698wr", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.54.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali44b9a32bd69", MAC:"96:35:ec:c2:3d:e5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:06.249004 containerd[1899]: 2026-03-02 12:50:06.182 [INFO][4833] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" Namespace="calico-system" Pod="whisker-cd5749c56-698wr" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-whisker--cd5749c56--698wr-eth0" Mar 2 12:50:07.438159 containerd[1899]: time="2026-03-02T12:50:07.437826844Z" level=info msg="connecting to shim 52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a" address="unix:///run/containerd/s/f1cf6ce54528e9aef3656909f73a4cc0ca9bf052c197d15cc3cd8d28b0617af9" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:07.455606 systemd[1]: Started cri-containerd-52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a.scope - libcontainer container 52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a. Mar 2 12:50:07.494083 containerd[1899]: time="2026-03-02T12:50:07.494007608Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-cd5749c56-698wr,Uid:3dfbabb8-d88d-426e-b3c4-32a46e087502,Namespace:calico-system,Attempt:0,} returns sandbox id \"52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a\"" Mar 2 12:50:07.495925 containerd[1899]: time="2026-03-02T12:50:07.495905880Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.3\"" Mar 2 12:50:07.666625 systemd-networkd[1485]: cali44b9a32bd69: Gained IPv6LL Mar 2 12:50:07.666859 systemd-networkd[1485]: vxlan.calico: Gained IPv6LL Mar 2 12:50:10.386512 containerd[1899]: time="2026-03-02T12:50:10.386401966Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:10.390806 containerd[1899]: time="2026-03-02T12:50:10.390782883Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.3: active requests=0, bytes read=5881068" Mar 2 12:50:10.435685 containerd[1899]: time="2026-03-02T12:50:10.435635244Z" level=info msg="ImageCreate event name:\"sha256:860a7f2cdb9123795f95a07e0cc91bc6b511927d1a4d1d588c303c9c59e0fa59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:10.483075 containerd[1899]: time="2026-03-02T12:50:10.483047365Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:3a388b567fff5cc31c64399d4af0fd03d2f4d243ef26e6f6b77a49386dbadeca\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:10.483879 containerd[1899]: time="2026-03-02T12:50:10.483807487Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.3\" with image id \"sha256:860a7f2cdb9123795f95a07e0cc91bc6b511927d1a4d1d588c303c9c59e0fa59\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:3a388b567fff5cc31c64399d4af0fd03d2f4d243ef26e6f6b77a49386dbadeca\", size \"7278585\" in 2.987792211s" Mar 2 12:50:10.483879 containerd[1899]: time="2026-03-02T12:50:10.483831224Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.3\" returns image reference \"sha256:860a7f2cdb9123795f95a07e0cc91bc6b511927d1a4d1d588c303c9c59e0fa59\"" Mar 2 12:50:10.532476 containerd[1899]: time="2026-03-02T12:50:10.532448130Z" level=info msg="CreateContainer within sandbox \"52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Mar 2 12:50:10.699870 containerd[1899]: time="2026-03-02T12:50:10.699534995Z" level=info msg="Container 4ea9ce2ce5d47ccd813c728c6dd7159942e2b14734a0bf1a14fd6cdf5caecc5a: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:10.700357 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount235229575.mount: Deactivated successfully. Mar 2 12:50:10.834315 containerd[1899]: time="2026-03-02T12:50:10.834274654Z" level=info msg="CreateContainer within sandbox \"52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"4ea9ce2ce5d47ccd813c728c6dd7159942e2b14734a0bf1a14fd6cdf5caecc5a\"" Mar 2 12:50:10.834763 containerd[1899]: time="2026-03-02T12:50:10.834738270Z" level=info msg="StartContainer for \"4ea9ce2ce5d47ccd813c728c6dd7159942e2b14734a0bf1a14fd6cdf5caecc5a\"" Mar 2 12:50:10.835465 containerd[1899]: time="2026-03-02T12:50:10.835436757Z" level=info msg="connecting to shim 4ea9ce2ce5d47ccd813c728c6dd7159942e2b14734a0bf1a14fd6cdf5caecc5a" address="unix:///run/containerd/s/f1cf6ce54528e9aef3656909f73a4cc0ca9bf052c197d15cc3cd8d28b0617af9" protocol=ttrpc version=3 Mar 2 12:50:10.858607 systemd[1]: Started cri-containerd-4ea9ce2ce5d47ccd813c728c6dd7159942e2b14734a0bf1a14fd6cdf5caecc5a.scope - libcontainer container 4ea9ce2ce5d47ccd813c728c6dd7159942e2b14734a0bf1a14fd6cdf5caecc5a. Mar 2 12:50:10.888683 containerd[1899]: time="2026-03-02T12:50:10.888653801Z" level=info msg="StartContainer for \"4ea9ce2ce5d47ccd813c728c6dd7159942e2b14734a0bf1a14fd6cdf5caecc5a\" returns successfully" Mar 2 12:50:10.890022 containerd[1899]: time="2026-03-02T12:50:10.889855353Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\"" Mar 2 12:50:12.388424 containerd[1899]: time="2026-03-02T12:50:12.388377625Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wtzgs,Uid:07af8044-862a-4c44-bd2c-12b2e3c44607,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:12.762283 systemd-networkd[1485]: cali34786afc5c8: Link UP Mar 2 12:50:12.762388 systemd-networkd[1485]: cali34786afc5c8: Gained carrier Mar 2 12:50:12.777818 containerd[1899]: 2026-03-02 12:50:12.709 [INFO][5020] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0 csi-node-driver- calico-system 07af8044-862a-4c44-bd2c-12b2e3c44607 729 0 2026-03-02 12:49:39 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:5d8f55657d k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 csi-node-driver-wtzgs eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali34786afc5c8 [] [] }} ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-" Mar 2 12:50:12.777818 containerd[1899]: 2026-03-02 12:50:12.709 [INFO][5020] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" Mar 2 12:50:12.777818 containerd[1899]: 2026-03-02 12:50:12.726 [INFO][5032] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" HandleID="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Workload="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.731 [INFO][5032] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" HandleID="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Workload="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fb4c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"csi-node-driver-wtzgs", "timestamp":"2026-03-02 12:50:12.726558532 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40003c7600)} Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.731 [INFO][5032] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.731 [INFO][5032] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.731 [INFO][5032] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.733 [INFO][5032] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.736 [INFO][5032] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.741 [INFO][5032] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.742 [INFO][5032] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.777976 containerd[1899]: 2026-03-02 12:50:12.744 [INFO][5032] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.778114 containerd[1899]: 2026-03-02 12:50:12.744 [INFO][5032] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.778114 containerd[1899]: 2026-03-02 12:50:12.745 [INFO][5032] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607 Mar 2 12:50:12.778114 containerd[1899]: 2026-03-02 12:50:12.752 [INFO][5032] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.778114 containerd[1899]: 2026-03-02 12:50:12.757 [INFO][5032] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.66/26] block=192.168.54.64/26 handle="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.778114 containerd[1899]: 2026-03-02 12:50:12.757 [INFO][5032] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.66/26] handle="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" host="ci-4459.2.101-47783670b7" Mar 2 12:50:12.778114 containerd[1899]: 2026-03-02 12:50:12.757 [INFO][5032] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:12.778114 containerd[1899]: 2026-03-02 12:50:12.757 [INFO][5032] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.66/26] IPv6=[] ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" HandleID="k8s-pod-network.c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Workload="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" Mar 2 12:50:12.778234 containerd[1899]: 2026-03-02 12:50:12.759 [INFO][5020] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"07af8044-862a-4c44-bd2c-12b2e3c44607", ResourceVersion:"729", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5d8f55657d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"csi-node-driver-wtzgs", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.54.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali34786afc5c8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:12.778270 containerd[1899]: 2026-03-02 12:50:12.759 [INFO][5020] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.66/32] ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" Mar 2 12:50:12.778270 containerd[1899]: 2026-03-02 12:50:12.759 [INFO][5020] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali34786afc5c8 ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" Mar 2 12:50:12.778270 containerd[1899]: 2026-03-02 12:50:12.761 [INFO][5020] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" Mar 2 12:50:12.778318 containerd[1899]: 2026-03-02 12:50:12.762 [INFO][5020] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"07af8044-862a-4c44-bd2c-12b2e3c44607", ResourceVersion:"729", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5d8f55657d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607", Pod:"csi-node-driver-wtzgs", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.54.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali34786afc5c8", MAC:"76:1a:0c:ac:ae:f1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:12.778351 containerd[1899]: 2026-03-02 12:50:12.772 [INFO][5020] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" Namespace="calico-system" Pod="csi-node-driver-wtzgs" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-csi--node--driver--wtzgs-eth0" Mar 2 12:50:13.094427 containerd[1899]: time="2026-03-02T12:50:13.094364836Z" level=info msg="connecting to shim c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607" address="unix:///run/containerd/s/b862d9a913fdd9bcc97e774726db3eba1b9272983c2647c18f376441208858ca" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:13.122639 systemd[1]: Started cri-containerd-c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607.scope - libcontainer container c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607. Mar 2 12:50:13.145066 containerd[1899]: time="2026-03-02T12:50:13.144992908Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wtzgs,Uid:07af8044-862a-4c44-bd2c-12b2e3c44607,Namespace:calico-system,Attempt:0,} returns sandbox id \"c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607\"" Mar 2 12:50:13.829273 containerd[1899]: time="2026-03-02T12:50:13.829232555Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6b7cf495fc-ftpl9,Uid:b7cc8850-f440-4499-b33a-69338a3f54c1,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:13.942925 containerd[1899]: time="2026-03-02T12:50:13.942777832Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rs7zg,Uid:7f11885f-bdcc-4b40-b0d0-d67b83c913b1,Namespace:kube-system,Attempt:0,}" Mar 2 12:50:13.988263 containerd[1899]: time="2026-03-02T12:50:13.988240058Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-p2qk6,Uid:9d4f30bf-d266-4607-b976-3eb09820fbc8,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:14.252772 systemd-networkd[1485]: cali00ac81965a8: Link UP Mar 2 12:50:14.253535 systemd-networkd[1485]: cali00ac81965a8: Gained carrier Mar 2 12:50:14.276345 containerd[1899]: 2026-03-02 12:50:14.129 [INFO][5105] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0 calico-kube-controllers-6b7cf495fc- calico-system b7cc8850-f440-4499-b33a-69338a3f54c1 876 0 2026-03-02 12:49:39 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6b7cf495fc projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 calico-kube-controllers-6b7cf495fc-ftpl9 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali00ac81965a8 [] [] }} ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-" Mar 2 12:50:14.276345 containerd[1899]: 2026-03-02 12:50:14.129 [INFO][5105] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" Mar 2 12:50:14.276345 containerd[1899]: 2026-03-02 12:50:14.152 [INFO][5121] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" HandleID="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Workload="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.187 [INFO][5121] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" HandleID="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Workload="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ed4b0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"calico-kube-controllers-6b7cf495fc-ftpl9", "timestamp":"2026-03-02 12:50:14.152487785 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400010c2c0)} Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.187 [INFO][5121] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.187 [INFO][5121] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.188 [INFO][5121] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.190 [INFO][5121] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.202 [INFO][5121] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.208 [INFO][5121] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.212 [INFO][5121] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276503 containerd[1899]: 2026-03-02 12:50:14.215 [INFO][5121] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276646 containerd[1899]: 2026-03-02 12:50:14.215 [INFO][5121] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276646 containerd[1899]: 2026-03-02 12:50:14.217 [INFO][5121] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2 Mar 2 12:50:14.276646 containerd[1899]: 2026-03-02 12:50:14.223 [INFO][5121] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276646 containerd[1899]: 2026-03-02 12:50:14.236 [INFO][5121] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.67/26] block=192.168.54.64/26 handle="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276646 containerd[1899]: 2026-03-02 12:50:14.236 [INFO][5121] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.67/26] handle="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.276646 containerd[1899]: 2026-03-02 12:50:14.236 [INFO][5121] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:14.276646 containerd[1899]: 2026-03-02 12:50:14.236 [INFO][5121] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.67/26] IPv6=[] ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" HandleID="k8s-pod-network.8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Workload="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" Mar 2 12:50:14.276740 containerd[1899]: 2026-03-02 12:50:14.244 [INFO][5105] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0", GenerateName:"calico-kube-controllers-6b7cf495fc-", Namespace:"calico-system", SelfLink:"", UID:"b7cc8850-f440-4499-b33a-69338a3f54c1", ResourceVersion:"876", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6b7cf495fc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"calico-kube-controllers-6b7cf495fc-ftpl9", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.54.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali00ac81965a8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:14.276775 containerd[1899]: 2026-03-02 12:50:14.244 [INFO][5105] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.67/32] ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" Mar 2 12:50:14.276775 containerd[1899]: 2026-03-02 12:50:14.244 [INFO][5105] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali00ac81965a8 ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" Mar 2 12:50:14.276775 containerd[1899]: 2026-03-02 12:50:14.253 [INFO][5105] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" Mar 2 12:50:14.276821 containerd[1899]: 2026-03-02 12:50:14.254 [INFO][5105] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0", GenerateName:"calico-kube-controllers-6b7cf495fc-", Namespace:"calico-system", SelfLink:"", UID:"b7cc8850-f440-4499-b33a-69338a3f54c1", ResourceVersion:"876", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6b7cf495fc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2", Pod:"calico-kube-controllers-6b7cf495fc-ftpl9", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.54.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali00ac81965a8", MAC:"3e:e6:29:72:2d:37", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:14.276855 containerd[1899]: 2026-03-02 12:50:14.271 [INFO][5105] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" Namespace="calico-system" Pod="calico-kube-controllers-6b7cf495fc-ftpl9" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--kube--controllers--6b7cf495fc--ftpl9-eth0" Mar 2 12:50:14.352201 systemd-networkd[1485]: cali992761394cb: Link UP Mar 2 12:50:14.353835 systemd-networkd[1485]: cali992761394cb: Gained carrier Mar 2 12:50:14.382723 containerd[1899]: 2026-03-02 12:50:14.240 [INFO][5127] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0 coredns-7d764666f9- kube-system 7f11885f-bdcc-4b40-b0d0-d67b83c913b1 874 0 2026-03-02 12:49:26 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7d764666f9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 coredns-7d764666f9-rs7zg eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali992761394cb [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-" Mar 2 12:50:14.382723 containerd[1899]: 2026-03-02 12:50:14.240 [INFO][5127] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" Mar 2 12:50:14.382723 containerd[1899]: 2026-03-02 12:50:14.286 [INFO][5150] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" HandleID="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Workload="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.295 [INFO][5150] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" HandleID="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Workload="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ed4b0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459.2.101-47783670b7", "pod":"coredns-7d764666f9-rs7zg", "timestamp":"2026-03-02 12:50:14.286215852 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40003ed080)} Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.295 [INFO][5150] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.295 [INFO][5150] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.295 [INFO][5150] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.298 [INFO][5150] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.304 [INFO][5150] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.314 [INFO][5150] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.319 [INFO][5150] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383410 containerd[1899]: 2026-03-02 12:50:14.321 [INFO][5150] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383598 containerd[1899]: 2026-03-02 12:50:14.321 [INFO][5150] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383598 containerd[1899]: 2026-03-02 12:50:14.322 [INFO][5150] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250 Mar 2 12:50:14.383598 containerd[1899]: 2026-03-02 12:50:14.331 [INFO][5150] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383598 containerd[1899]: 2026-03-02 12:50:14.341 [INFO][5150] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.68/26] block=192.168.54.64/26 handle="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383598 containerd[1899]: 2026-03-02 12:50:14.341 [INFO][5150] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.68/26] handle="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.383598 containerd[1899]: 2026-03-02 12:50:14.341 [INFO][5150] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:14.383598 containerd[1899]: 2026-03-02 12:50:14.341 [INFO][5150] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.68/26] IPv6=[] ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" HandleID="k8s-pod-network.443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Workload="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" Mar 2 12:50:14.384949 containerd[1899]: 2026-03-02 12:50:14.345 [INFO][5127] cni-plugin/k8s.go 418: Populated endpoint ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"7f11885f-bdcc-4b40-b0d0-d67b83c913b1", ResourceVersion:"874", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"coredns-7d764666f9-rs7zg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.54.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali992761394cb", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:14.384949 containerd[1899]: 2026-03-02 12:50:14.345 [INFO][5127] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.68/32] ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" Mar 2 12:50:14.384949 containerd[1899]: 2026-03-02 12:50:14.345 [INFO][5127] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali992761394cb ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" Mar 2 12:50:14.384949 containerd[1899]: 2026-03-02 12:50:14.359 [INFO][5127] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" Mar 2 12:50:14.384949 containerd[1899]: 2026-03-02 12:50:14.360 [INFO][5127] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"7f11885f-bdcc-4b40-b0d0-d67b83c913b1", ResourceVersion:"874", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250", Pod:"coredns-7d764666f9-rs7zg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.54.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali992761394cb", MAC:"d6:5c:b0:9a:7b:87", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:14.385087 containerd[1899]: 2026-03-02 12:50:14.378 [INFO][5127] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" Namespace="kube-system" Pod="coredns-7d764666f9-rs7zg" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--rs7zg-eth0" Mar 2 12:50:14.398546 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount630362435.mount: Deactivated successfully. Mar 2 12:50:14.437586 systemd-networkd[1485]: cali01cb78abbad: Link UP Mar 2 12:50:14.438485 systemd-networkd[1485]: cali01cb78abbad: Gained carrier Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.304 [INFO][5141] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0 calico-apiserver-8487494cdb- calico-system 9d4f30bf-d266-4607-b976-3eb09820fbc8 875 0 2026-03-02 12:49:38 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:8487494cdb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 calico-apiserver-8487494cdb-p2qk6 eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali01cb78abbad [] [] }} ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.304 [INFO][5141] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.344 [INFO][5169] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.354 [INFO][5169] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002f3a50), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"calico-apiserver-8487494cdb-p2qk6", "timestamp":"2026-03-02 12:50:14.344451268 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400027af20)} Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.355 [INFO][5169] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.355 [INFO][5169] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.355 [INFO][5169] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.399 [INFO][5169] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.404 [INFO][5169] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.410 [INFO][5169] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.411 [INFO][5169] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.413 [INFO][5169] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.413 [INFO][5169] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.414 [INFO][5169] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423 Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.420 [INFO][5169] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.431 [INFO][5169] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.69/26] block=192.168.54.64/26 handle="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.432 [INFO][5169] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.69/26] handle="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" host="ci-4459.2.101-47783670b7" Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.432 [INFO][5169] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:14.451034 containerd[1899]: 2026-03-02 12:50:14.432 [INFO][5169] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.69/26] IPv6=[] ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:14.451470 containerd[1899]: 2026-03-02 12:50:14.435 [INFO][5141] cni-plugin/k8s.go 418: Populated endpoint ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0", GenerateName:"calico-apiserver-8487494cdb-", Namespace:"calico-system", SelfLink:"", UID:"9d4f30bf-d266-4607-b976-3eb09820fbc8", ResourceVersion:"875", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8487494cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"calico-apiserver-8487494cdb-p2qk6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali01cb78abbad", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:14.451470 containerd[1899]: 2026-03-02 12:50:14.435 [INFO][5141] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.69/32] ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:14.451470 containerd[1899]: 2026-03-02 12:50:14.435 [INFO][5141] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali01cb78abbad ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:14.451470 containerd[1899]: 2026-03-02 12:50:14.439 [INFO][5141] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:14.451470 containerd[1899]: 2026-03-02 12:50:14.439 [INFO][5141] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0", GenerateName:"calico-apiserver-8487494cdb-", Namespace:"calico-system", SelfLink:"", UID:"9d4f30bf-d266-4607-b976-3eb09820fbc8", ResourceVersion:"875", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8487494cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423", Pod:"calico-apiserver-8487494cdb-p2qk6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali01cb78abbad", MAC:"1e:2d:36:d8:f7:b6", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:14.451470 containerd[1899]: 2026-03-02 12:50:14.448 [INFO][5141] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-p2qk6" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:14.514594 systemd-networkd[1485]: cali34786afc5c8: Gained IPv6LL Mar 2 12:50:14.740614 containerd[1899]: time="2026-03-02T12:50:14.740572852Z" level=info msg="connecting to shim 8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2" address="unix:///run/containerd/s/52dd938832a1047e158113ed957b20895882146d45e92b5dc8f59e062af6886c" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:14.755600 systemd[1]: Started cri-containerd-8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2.scope - libcontainer container 8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2. Mar 2 12:50:14.986407 containerd[1899]: time="2026-03-02T12:50:14.986357579Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6b7cf495fc-ftpl9,Uid:b7cc8850-f440-4499-b33a-69338a3f54c1,Namespace:calico-system,Attempt:0,} returns sandbox id \"8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2\"" Mar 2 12:50:15.387119 containerd[1899]: time="2026-03-02T12:50:15.387080838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-g2d5q,Uid:f9e26872-e431-4fd3-808e-5bafbba0261e,Namespace:kube-system,Attempt:0,}" Mar 2 12:50:15.434196 containerd[1899]: time="2026-03-02T12:50:15.434125469Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-7zh2j,Uid:56b6d31a-681f-4836-a653-77f50d324a86,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:15.667034 systemd-networkd[1485]: cali992761394cb: Gained IPv6LL Mar 2 12:50:15.696963 containerd[1899]: time="2026-03-02T12:50:15.696661527Z" level=info msg="connecting to shim 443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250" address="unix:///run/containerd/s/d55b90310d475ce59985ee74824ecafdb63fc4eef32fc11447eb46a61218f06d" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:15.722608 systemd[1]: Started cri-containerd-443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250.scope - libcontainer container 443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250. Mar 2 12:50:15.848565 containerd[1899]: time="2026-03-02T12:50:15.848540013Z" level=info msg="connecting to shim bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" address="unix:///run/containerd/s/48219788bc27c5f50459db8160775c0ceede6cffe029185097fe9d92baf7dcd6" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:15.859278 systemd-networkd[1485]: cali01cb78abbad: Gained IPv6LL Mar 2 12:50:15.863623 systemd[1]: Started cri-containerd-bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423.scope - libcontainer container bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423. Mar 2 12:50:15.938219 containerd[1899]: time="2026-03-02T12:50:15.938132356Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-rs7zg,Uid:7f11885f-bdcc-4b40-b0d0-d67b83c913b1,Namespace:kube-system,Attempt:0,} returns sandbox id \"443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250\"" Mar 2 12:50:15.980197 systemd-networkd[1485]: cali71b08fa4332: Link UP Mar 2 12:50:15.981206 systemd-networkd[1485]: cali71b08fa4332: Gained carrier Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.919 [INFO][5383] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0 coredns-7d764666f9- kube-system f9e26872-e431-4fd3-808e-5bafbba0261e 873 0 2026-03-02 12:49:26 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7d764666f9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 coredns-7d764666f9-g2d5q eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali71b08fa4332 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.920 [INFO][5383] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.938 [INFO][5395] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" HandleID="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Workload="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.944 [INFO][5395] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" HandleID="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Workload="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fbe80), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4459.2.101-47783670b7", "pod":"coredns-7d764666f9-g2d5q", "timestamp":"2026-03-02 12:50:15.93802088 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000186dc0)} Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.944 [INFO][5395] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.944 [INFO][5395] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.944 [INFO][5395] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.950 [INFO][5395] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.953 [INFO][5395] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.958 [INFO][5395] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.959 [INFO][5395] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.961 [INFO][5395] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.961 [INFO][5395] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.962 [INFO][5395] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.966 [INFO][5395] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.975 [INFO][5395] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.70/26] block=192.168.54.64/26 handle="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.975 [INFO][5395] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.70/26] handle="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.975 [INFO][5395] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:16.033424 containerd[1899]: 2026-03-02 12:50:15.975 [INFO][5395] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.70/26] IPv6=[] ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" HandleID="k8s-pod-network.1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Workload="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" Mar 2 12:50:16.034261 containerd[1899]: 2026-03-02 12:50:15.977 [INFO][5383] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"f9e26872-e431-4fd3-808e-5bafbba0261e", ResourceVersion:"873", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"coredns-7d764666f9-g2d5q", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.54.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali71b08fa4332", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:16.034261 containerd[1899]: 2026-03-02 12:50:15.977 [INFO][5383] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.70/32] ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" Mar 2 12:50:16.034261 containerd[1899]: 2026-03-02 12:50:15.977 [INFO][5383] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali71b08fa4332 ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" Mar 2 12:50:16.034261 containerd[1899]: 2026-03-02 12:50:15.981 [INFO][5383] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" Mar 2 12:50:16.034261 containerd[1899]: 2026-03-02 12:50:15.982 [INFO][5383] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0", GenerateName:"coredns-7d764666f9-", Namespace:"kube-system", SelfLink:"", UID:"f9e26872-e431-4fd3-808e-5bafbba0261e", ResourceVersion:"873", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7d764666f9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b", Pod:"coredns-7d764666f9-g2d5q", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.54.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali71b08fa4332", MAC:"62:06:c8:ca:09:e0", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:16.034392 containerd[1899]: 2026-03-02 12:50:15.995 [INFO][5383] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" Namespace="kube-system" Pod="coredns-7d764666f9-g2d5q" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-coredns--7d764666f9--g2d5q-eth0" Mar 2 12:50:16.091446 containerd[1899]: time="2026-03-02T12:50:16.091409217Z" level=info msg="CreateContainer within sandbox \"443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 12:50:16.094333 containerd[1899]: time="2026-03-02T12:50:16.094292842Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-p2qk6,Uid:9d4f30bf-d266-4607-b976-3eb09820fbc8,Namespace:calico-system,Attempt:0,} returns sandbox id \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\"" Mar 2 12:50:16.104475 systemd-networkd[1485]: cali3b23f72654d: Link UP Mar 2 12:50:16.105682 systemd-networkd[1485]: cali3b23f72654d: Gained carrier Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.018 [INFO][5403] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0 calico-apiserver-8487494cdb- calico-system 56b6d31a-681f-4836-a653-77f50d324a86 880 0 2026-03-02 12:49:38 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:8487494cdb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 calico-apiserver-8487494cdb-7zh2j eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali3b23f72654d [] [] }} ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.031 [INFO][5403] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.051 [INFO][5429] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.056 [INFO][5429] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000273350), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"calico-apiserver-8487494cdb-7zh2j", "timestamp":"2026-03-02 12:50:16.051426752 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400038ef20)} Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.056 [INFO][5429] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.056 [INFO][5429] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.056 [INFO][5429] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.058 [INFO][5429] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.061 [INFO][5429] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.063 [INFO][5429] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.065 [INFO][5429] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.066 [INFO][5429] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.066 [INFO][5429] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.067 [INFO][5429] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51 Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.083 [INFO][5429] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.096 [INFO][5429] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.71/26] block=192.168.54.64/26 handle="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.096 [INFO][5429] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.71/26] handle="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.096 [INFO][5429] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:16.133341 containerd[1899]: 2026-03-02 12:50:16.096 [INFO][5429] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.71/26] IPv6=[] ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:16.133724 containerd[1899]: 2026-03-02 12:50:16.100 [INFO][5403] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0", GenerateName:"calico-apiserver-8487494cdb-", Namespace:"calico-system", SelfLink:"", UID:"56b6d31a-681f-4836-a653-77f50d324a86", ResourceVersion:"880", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8487494cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"calico-apiserver-8487494cdb-7zh2j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali3b23f72654d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:16.133724 containerd[1899]: 2026-03-02 12:50:16.100 [INFO][5403] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.71/32] ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:16.133724 containerd[1899]: 2026-03-02 12:50:16.100 [INFO][5403] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3b23f72654d ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:16.133724 containerd[1899]: 2026-03-02 12:50:16.106 [INFO][5403] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:16.133724 containerd[1899]: 2026-03-02 12:50:16.109 [INFO][5403] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0", GenerateName:"calico-apiserver-8487494cdb-", Namespace:"calico-system", SelfLink:"", UID:"56b6d31a-681f-4836-a653-77f50d324a86", ResourceVersion:"880", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8487494cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51", Pod:"calico-apiserver-8487494cdb-7zh2j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali3b23f72654d", MAC:"56:81:14:65:fa:0a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:16.133724 containerd[1899]: 2026-03-02 12:50:16.130 [INFO][5403] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Namespace="calico-system" Pod="calico-apiserver-8487494cdb-7zh2j" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:16.178590 systemd-networkd[1485]: cali00ac81965a8: Gained IPv6LL Mar 2 12:50:16.189904 containerd[1899]: time="2026-03-02T12:50:16.189818977Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:16.335374 containerd[1899]: time="2026-03-02T12:50:16.335339497Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.3: active requests=0, bytes read=16420592" Mar 2 12:50:16.490541 containerd[1899]: time="2026-03-02T12:50:16.490440156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7d7658d587-6g28g,Uid:cfb890bf-40cf-4b4a-b482-8dad57b17735,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:16.585393 containerd[1899]: time="2026-03-02T12:50:16.585329909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68bfcfb4bc-jjtz2,Uid:a6e52250-759f-484f-86cd-140ce9b02f00,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:16.634812 containerd[1899]: time="2026-03-02T12:50:16.634784821Z" level=info msg="ImageCreate event name:\"sha256:d6c2d25ea514599ef2dbba86e46277491ee9c1e15519321c135bb514b2f46aeb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:16.974627 systemd-networkd[1485]: califb7dc352b22: Link UP Mar 2 12:50:16.977111 systemd-networkd[1485]: califb7dc352b22: Gained carrier Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.908 [INFO][5484] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0 goldmane-7d7658d587- calico-system cfb890bf-40cf-4b4a-b482-8dad57b17735 879 0 2026-03-02 12:49:38 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:7d7658d587 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 goldmane-7d7658d587-6g28g eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] califb7dc352b22 [] [] }} ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.909 [INFO][5484] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.928 [INFO][5497] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" HandleID="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Workload="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.934 [INFO][5497] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" HandleID="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Workload="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000273220), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"goldmane-7d7658d587-6g28g", "timestamp":"2026-03-02 12:50:16.928472079 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40002e51e0)} Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.934 [INFO][5497] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.934 [INFO][5497] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.934 [INFO][5497] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.936 [INFO][5497] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.942 [INFO][5497] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.952 [INFO][5497] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.953 [INFO][5497] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.954 [INFO][5497] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.954 [INFO][5497] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.956 [INFO][5497] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.959 [INFO][5497] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.967 [INFO][5497] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.72/26] block=192.168.54.64/26 handle="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.968 [INFO][5497] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.72/26] handle="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" host="ci-4459.2.101-47783670b7" Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.968 [INFO][5497] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:16.993170 containerd[1899]: 2026-03-02 12:50:16.968 [INFO][5497] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.72/26] IPv6=[] ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" HandleID="k8s-pod-network.68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Workload="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" Mar 2 12:50:16.995058 containerd[1899]: 2026-03-02 12:50:16.970 [INFO][5484] cni-plugin/k8s.go 418: Populated endpoint ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0", GenerateName:"goldmane-7d7658d587-", Namespace:"calico-system", SelfLink:"", UID:"cfb890bf-40cf-4b4a-b482-8dad57b17735", ResourceVersion:"879", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7d7658d587", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"goldmane-7d7658d587-6g28g", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.54.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"califb7dc352b22", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:16.995058 containerd[1899]: 2026-03-02 12:50:16.970 [INFO][5484] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.72/32] ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" Mar 2 12:50:16.995058 containerd[1899]: 2026-03-02 12:50:16.970 [INFO][5484] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to califb7dc352b22 ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" Mar 2 12:50:16.995058 containerd[1899]: 2026-03-02 12:50:16.974 [INFO][5484] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" Mar 2 12:50:16.995058 containerd[1899]: 2026-03-02 12:50:16.976 [INFO][5484] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0", GenerateName:"goldmane-7d7658d587-", Namespace:"calico-system", SelfLink:"", UID:"cfb890bf-40cf-4b4a-b482-8dad57b17735", ResourceVersion:"879", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7d7658d587", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed", Pod:"goldmane-7d7658d587-6g28g", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.54.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"califb7dc352b22", MAC:"96:f0:ed:b1:b9:3a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:16.995058 containerd[1899]: 2026-03-02 12:50:16.988 [INFO][5484] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" Namespace="calico-system" Pod="goldmane-7d7658d587-6g28g" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-goldmane--7d7658d587--6g28g-eth0" Mar 2 12:50:17.001982 containerd[1899]: time="2026-03-02T12:50:17.000145411Z" level=info msg="Container c0f42a123a9c5ae43f3b114df398177a9db7b25665de6a0e532d9a6cfa233874: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:17.000357 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2625218745.mount: Deactivated successfully. Mar 2 12:50:17.202578 systemd-networkd[1485]: cali3b23f72654d: Gained IPv6LL Mar 2 12:50:17.330237 containerd[1899]: time="2026-03-02T12:50:17.330136523Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:359cb5c751e049ac0bb62c4f7e49b1ac81c59935c70715f5ff4c39a757bf9f38\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:17.333126 containerd[1899]: time="2026-03-02T12:50:17.332974058Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\" with image id \"sha256:d6c2d25ea514599ef2dbba86e46277491ee9c1e15519321c135bb514b2f46aeb\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:359cb5c751e049ac0bb62c4f7e49b1ac81c59935c70715f5ff4c39a757bf9f38\", size \"16420422\" in 6.443090536s" Mar 2 12:50:17.333126 containerd[1899]: time="2026-03-02T12:50:17.333095046Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.3\" returns image reference \"sha256:d6c2d25ea514599ef2dbba86e46277491ee9c1e15519321c135bb514b2f46aeb\"" Mar 2 12:50:17.391513 containerd[1899]: time="2026-03-02T12:50:17.335413364Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.3\"" Mar 2 12:50:17.432143 systemd-networkd[1485]: cali156885eb2ce: Link UP Mar 2 12:50:17.432244 systemd-networkd[1485]: cali156885eb2ce: Gained carrier Mar 2 12:50:17.441837 containerd[1899]: time="2026-03-02T12:50:17.441776903Z" level=info msg="CreateContainer within sandbox \"52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.307 [INFO][5534] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0 calico-apiserver-68bfcfb4bc- calico-system a6e52250-759f-484f-86cd-140ce9b02f00 877 0 2026-03-02 12:49:39 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:68bfcfb4bc projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 calico-apiserver-68bfcfb4bc-jjtz2 eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali156885eb2ce [] [] }} ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.307 [INFO][5534] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.323 [INFO][5547] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" HandleID="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.328 [INFO][5547] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" HandleID="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ed4b0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"calico-apiserver-68bfcfb4bc-jjtz2", "timestamp":"2026-03-02 12:50:17.323747836 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40003b5080)} Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.329 [INFO][5547] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.329 [INFO][5547] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.329 [INFO][5547] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.331 [INFO][5547] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.336 [INFO][5547] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.395 [INFO][5547] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.396 [INFO][5547] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.398 [INFO][5547] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.398 [INFO][5547] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.399 [INFO][5547] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0 Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.405 [INFO][5547] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.427 [INFO][5547] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.73/26] block=192.168.54.64/26 handle="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.428 [INFO][5547] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.73/26] handle="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" host="ci-4459.2.101-47783670b7" Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.428 [INFO][5547] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:17.446958 containerd[1899]: 2026-03-02 12:50:17.428 [INFO][5547] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.73/26] IPv6=[] ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" HandleID="k8s-pod-network.dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" Mar 2 12:50:17.447307 containerd[1899]: 2026-03-02 12:50:17.430 [INFO][5534] cni-plugin/k8s.go 418: Populated endpoint ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0", GenerateName:"calico-apiserver-68bfcfb4bc-", Namespace:"calico-system", SelfLink:"", UID:"a6e52250-759f-484f-86cd-140ce9b02f00", ResourceVersion:"877", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68bfcfb4bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"calico-apiserver-68bfcfb4bc-jjtz2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali156885eb2ce", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:17.447307 containerd[1899]: 2026-03-02 12:50:17.430 [INFO][5534] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.73/32] ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" Mar 2 12:50:17.447307 containerd[1899]: 2026-03-02 12:50:17.430 [INFO][5534] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali156885eb2ce ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" Mar 2 12:50:17.447307 containerd[1899]: 2026-03-02 12:50:17.432 [INFO][5534] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" Mar 2 12:50:17.447307 containerd[1899]: 2026-03-02 12:50:17.432 [INFO][5534] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0", GenerateName:"calico-apiserver-68bfcfb4bc-", Namespace:"calico-system", SelfLink:"", UID:"a6e52250-759f-484f-86cd-140ce9b02f00", ResourceVersion:"877", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 49, 39, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68bfcfb4bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0", Pod:"calico-apiserver-68bfcfb4bc-jjtz2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali156885eb2ce", MAC:"3e:a8:97:da:15:1e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:17.447307 containerd[1899]: 2026-03-02 12:50:17.443 [INFO][5534] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-jjtz2" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--jjtz2-eth0" Mar 2 12:50:17.458595 systemd-networkd[1485]: cali71b08fa4332: Gained IPv6LL Mar 2 12:50:17.591820 containerd[1899]: time="2026-03-02T12:50:17.591788255Z" level=info msg="connecting to shim ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" address="unix:///run/containerd/s/6043b40311a0eac475e7477ae43eda7d8dca1c7b3475c4502f9a62f33f52c990" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:17.611667 systemd[1]: Started cri-containerd-ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51.scope - libcontainer container ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51. Mar 2 12:50:18.098641 systemd-networkd[1485]: califb7dc352b22: Gained IPv6LL Mar 2 12:50:18.357975 containerd[1899]: time="2026-03-02T12:50:18.357861800Z" level=info msg="CreateContainer within sandbox \"443f1eadbd53f559428330e71e1bac75f12cdbe3d0312eec17fab6604d0fe250\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c0f42a123a9c5ae43f3b114df398177a9db7b25665de6a0e532d9a6cfa233874\"" Mar 2 12:50:18.359216 containerd[1899]: time="2026-03-02T12:50:18.359187933Z" level=info msg="StartContainer for \"c0f42a123a9c5ae43f3b114df398177a9db7b25665de6a0e532d9a6cfa233874\"" Mar 2 12:50:18.359829 containerd[1899]: time="2026-03-02T12:50:18.359806178Z" level=info msg="connecting to shim c0f42a123a9c5ae43f3b114df398177a9db7b25665de6a0e532d9a6cfa233874" address="unix:///run/containerd/s/d55b90310d475ce59985ee74824ecafdb63fc4eef32fc11447eb46a61218f06d" protocol=ttrpc version=3 Mar 2 12:50:18.382333 containerd[1899]: time="2026-03-02T12:50:18.382301279Z" level=info msg="connecting to shim 1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b" address="unix:///run/containerd/s/ca6847dd81342f02215ed57ca160169efd0eaa7da5fcd3000ea77bb843b51cda" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:18.386626 systemd[1]: Started cri-containerd-c0f42a123a9c5ae43f3b114df398177a9db7b25665de6a0e532d9a6cfa233874.scope - libcontainer container c0f42a123a9c5ae43f3b114df398177a9db7b25665de6a0e532d9a6cfa233874. Mar 2 12:50:18.392550 containerd[1899]: time="2026-03-02T12:50:18.392353409Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8487494cdb-7zh2j,Uid:56b6d31a-681f-4836-a653-77f50d324a86,Namespace:calico-system,Attempt:0,} returns sandbox id \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\"" Mar 2 12:50:18.408823 systemd[1]: Started cri-containerd-1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b.scope - libcontainer container 1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b. Mar 2 12:50:18.425325 containerd[1899]: time="2026-03-02T12:50:18.425263628Z" level=info msg="Container abcf9b630ccaefd8e717138fb4021a084aa19fbc62ec4c8279dbd3b75040a7f1: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:18.438998 containerd[1899]: time="2026-03-02T12:50:18.438884423Z" level=info msg="StartContainer for \"c0f42a123a9c5ae43f3b114df398177a9db7b25665de6a0e532d9a6cfa233874\" returns successfully" Mar 2 12:50:18.445624 containerd[1899]: time="2026-03-02T12:50:18.445599449Z" level=info msg="connecting to shim 68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed" address="unix:///run/containerd/s/426942d99f1e521528ecfdaf82b5014db3bd9be5db92197c3393535aa40933d0" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:18.469027 containerd[1899]: time="2026-03-02T12:50:18.468999940Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-g2d5q,Uid:f9e26872-e431-4fd3-808e-5bafbba0261e,Namespace:kube-system,Attempt:0,} returns sandbox id \"1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b\"" Mar 2 12:50:18.470813 systemd[1]: Started cri-containerd-68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed.scope - libcontainer container 68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed. Mar 2 12:50:18.471819 containerd[1899]: time="2026-03-02T12:50:18.471098883Z" level=info msg="CreateContainer within sandbox \"52649f8a827355078cc8c4933503630d677be2f82bb8ed92767676de682d831a\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"abcf9b630ccaefd8e717138fb4021a084aa19fbc62ec4c8279dbd3b75040a7f1\"" Mar 2 12:50:18.473614 containerd[1899]: time="2026-03-02T12:50:18.472668335Z" level=info msg="StartContainer for \"abcf9b630ccaefd8e717138fb4021a084aa19fbc62ec4c8279dbd3b75040a7f1\"" Mar 2 12:50:18.475632 containerd[1899]: time="2026-03-02T12:50:18.475605434Z" level=info msg="connecting to shim abcf9b630ccaefd8e717138fb4021a084aa19fbc62ec4c8279dbd3b75040a7f1" address="unix:///run/containerd/s/f1cf6ce54528e9aef3656909f73a4cc0ca9bf052c197d15cc3cd8d28b0617af9" protocol=ttrpc version=3 Mar 2 12:50:18.481747 containerd[1899]: time="2026-03-02T12:50:18.481623829Z" level=info msg="CreateContainer within sandbox \"1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 12:50:18.495930 containerd[1899]: time="2026-03-02T12:50:18.495905357Z" level=info msg="connecting to shim dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0" address="unix:///run/containerd/s/f793bff982e5b0e95978a8cedbdd3783d5985c138deb1b3178629b5cf231feb5" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:18.507666 systemd[1]: Started cri-containerd-abcf9b630ccaefd8e717138fb4021a084aa19fbc62ec4c8279dbd3b75040a7f1.scope - libcontainer container abcf9b630ccaefd8e717138fb4021a084aa19fbc62ec4c8279dbd3b75040a7f1. Mar 2 12:50:18.511416 containerd[1899]: time="2026-03-02T12:50:18.511192168Z" level=info msg="Container 646eb2b05ea8f24561d64d7439f8b07826b8f2dea1aa72769fdc927e4749714f: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:18.526623 systemd[1]: Started cri-containerd-dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0.scope - libcontainer container dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0. Mar 2 12:50:18.539950 containerd[1899]: time="2026-03-02T12:50:18.539827627Z" level=info msg="CreateContainer within sandbox \"1be2bca9b25355671d59bb4087a986afab9c8d691a2ede92e01a0453efa45e6b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"646eb2b05ea8f24561d64d7439f8b07826b8f2dea1aa72769fdc927e4749714f\"" Mar 2 12:50:18.544761 containerd[1899]: time="2026-03-02T12:50:18.543659308Z" level=info msg="StartContainer for \"646eb2b05ea8f24561d64d7439f8b07826b8f2dea1aa72769fdc927e4749714f\"" Mar 2 12:50:18.550972 containerd[1899]: time="2026-03-02T12:50:18.550950105Z" level=info msg="connecting to shim 646eb2b05ea8f24561d64d7439f8b07826b8f2dea1aa72769fdc927e4749714f" address="unix:///run/containerd/s/ca6847dd81342f02215ed57ca160169efd0eaa7da5fcd3000ea77bb843b51cda" protocol=ttrpc version=3 Mar 2 12:50:18.559746 containerd[1899]: time="2026-03-02T12:50:18.558399444Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7d7658d587-6g28g,Uid:cfb890bf-40cf-4b4a-b482-8dad57b17735,Namespace:calico-system,Attempt:0,} returns sandbox id \"68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed\"" Mar 2 12:50:18.588824 systemd[1]: Started cri-containerd-646eb2b05ea8f24561d64d7439f8b07826b8f2dea1aa72769fdc927e4749714f.scope - libcontainer container 646eb2b05ea8f24561d64d7439f8b07826b8f2dea1aa72769fdc927e4749714f. Mar 2 12:50:18.592381 kubelet[3475]: I0302 12:50:18.592322 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-rs7zg" podStartSLOduration=52.592309417 podStartE2EDuration="52.592309417s" podCreationTimestamp="2026-03-02 12:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:50:18.591561016 +0000 UTC m=+59.293891795" watchObservedRunningTime="2026-03-02 12:50:18.592309417 +0000 UTC m=+59.294640180" Mar 2 12:50:18.597386 containerd[1899]: time="2026-03-02T12:50:18.597350075Z" level=info msg="StartContainer for \"abcf9b630ccaefd8e717138fb4021a084aa19fbc62ec4c8279dbd3b75040a7f1\" returns successfully" Mar 2 12:50:18.605598 containerd[1899]: time="2026-03-02T12:50:18.605572511Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68bfcfb4bc-jjtz2,Uid:a6e52250-759f-484f-86cd-140ce9b02f00,Namespace:calico-system,Attempt:0,} returns sandbox id \"dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0\"" Mar 2 12:50:18.654653 containerd[1899]: time="2026-03-02T12:50:18.649642442Z" level=info msg="StartContainer for \"646eb2b05ea8f24561d64d7439f8b07826b8f2dea1aa72769fdc927e4749714f\" returns successfully" Mar 2 12:50:18.942284 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2840310071.mount: Deactivated successfully. Mar 2 12:50:19.058628 systemd-networkd[1485]: cali156885eb2ce: Gained IPv6LL Mar 2 12:50:19.591228 kubelet[3475]: I0302 12:50:19.591172 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-g2d5q" podStartSLOduration=53.59115791 podStartE2EDuration="53.59115791s" podCreationTimestamp="2026-03-02 12:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:50:19.590660989 +0000 UTC m=+60.292991728" watchObservedRunningTime="2026-03-02 12:50:19.59115791 +0000 UTC m=+60.293488657" Mar 2 12:50:19.826632 containerd[1899]: time="2026-03-02T12:50:19.826584816Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:19.829696 containerd[1899]: time="2026-03-02T12:50:19.829669752Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.3: active requests=0, bytes read=8255947" Mar 2 12:50:19.833069 containerd[1899]: time="2026-03-02T12:50:19.833041113Z" level=info msg="ImageCreate event name:\"sha256:a7b37b6d011a8219915c610022e2c5ef47396285db6e7e10d7694ff3dea87dc5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:19.837886 containerd[1899]: time="2026-03-02T12:50:19.837858267Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:3d04cd6265f850f0420b413351275ebfd244991b1b9e69c64efe8b4eff45b53f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:19.840096 containerd[1899]: time="2026-03-02T12:50:19.840051045Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.3\" with image id \"sha256:a7b37b6d011a8219915c610022e2c5ef47396285db6e7e10d7694ff3dea87dc5\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:3d04cd6265f850f0420b413351275ebfd244991b1b9e69c64efe8b4eff45b53f\", size \"9653472\" in 2.504608776s" Mar 2 12:50:19.840096 containerd[1899]: time="2026-03-02T12:50:19.840088702Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.3\" returns image reference \"sha256:a7b37b6d011a8219915c610022e2c5ef47396285db6e7e10d7694ff3dea87dc5\"" Mar 2 12:50:19.840884 containerd[1899]: time="2026-03-02T12:50:19.840803310Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\"" Mar 2 12:50:19.849349 containerd[1899]: time="2026-03-02T12:50:19.849280475Z" level=info msg="CreateContainer within sandbox \"c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 2 12:50:19.876825 containerd[1899]: time="2026-03-02T12:50:19.876797255Z" level=info msg="Container 2b9ad462f211b101d9ba166c30f4fb90479a346fa6a056f17a310e1d50fe56a7: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:19.893925 containerd[1899]: time="2026-03-02T12:50:19.893879597Z" level=info msg="CreateContainer within sandbox \"c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"2b9ad462f211b101d9ba166c30f4fb90479a346fa6a056f17a310e1d50fe56a7\"" Mar 2 12:50:19.895539 containerd[1899]: time="2026-03-02T12:50:19.894807156Z" level=info msg="StartContainer for \"2b9ad462f211b101d9ba166c30f4fb90479a346fa6a056f17a310e1d50fe56a7\"" Mar 2 12:50:19.896404 containerd[1899]: time="2026-03-02T12:50:19.896377137Z" level=info msg="connecting to shim 2b9ad462f211b101d9ba166c30f4fb90479a346fa6a056f17a310e1d50fe56a7" address="unix:///run/containerd/s/b862d9a913fdd9bcc97e774726db3eba1b9272983c2647c18f376441208858ca" protocol=ttrpc version=3 Mar 2 12:50:19.912613 systemd[1]: Started cri-containerd-2b9ad462f211b101d9ba166c30f4fb90479a346fa6a056f17a310e1d50fe56a7.scope - libcontainer container 2b9ad462f211b101d9ba166c30f4fb90479a346fa6a056f17a310e1d50fe56a7. Mar 2 12:50:19.964873 containerd[1899]: time="2026-03-02T12:50:19.964845596Z" level=info msg="StartContainer for \"2b9ad462f211b101d9ba166c30f4fb90479a346fa6a056f17a310e1d50fe56a7\" returns successfully" Mar 2 12:50:22.733131 containerd[1899]: time="2026-03-02T12:50:22.733081796Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:22.741495 containerd[1899]: time="2026-03-02T12:50:22.741460045Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.3: active requests=0, bytes read=49157508" Mar 2 12:50:22.749061 containerd[1899]: time="2026-03-02T12:50:22.749032179Z" level=info msg="ImageCreate event name:\"sha256:f91182157dd9b43afadc3f9d6dbd919b0ec222fc40e9fa608989310b81c1f18c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:22.755670 containerd[1899]: time="2026-03-02T12:50:22.755626633Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:081fd6c3de7754ba9892532b2c7c6cae9ba7bd1cca4c42e4590ee8d0f5a5696b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:22.756259 containerd[1899]: time="2026-03-02T12:50:22.755911491Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\" with image id \"sha256:f91182157dd9b43afadc3f9d6dbd919b0ec222fc40e9fa608989310b81c1f18c\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:081fd6c3de7754ba9892532b2c7c6cae9ba7bd1cca4c42e4590ee8d0f5a5696b\", size \"50555001\" in 2.915058594s" Mar 2 12:50:22.756259 containerd[1899]: time="2026-03-02T12:50:22.755937883Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.3\" returns image reference \"sha256:f91182157dd9b43afadc3f9d6dbd919b0ec222fc40e9fa608989310b81c1f18c\"" Mar 2 12:50:22.757147 containerd[1899]: time="2026-03-02T12:50:22.756923476Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\"" Mar 2 12:50:22.778704 containerd[1899]: time="2026-03-02T12:50:22.778667767Z" level=info msg="CreateContainer within sandbox \"8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 2 12:50:22.800966 containerd[1899]: time="2026-03-02T12:50:22.800629848Z" level=info msg="Container 1c856bf3d53ce0fd79e61b57cd6663c8ccce85378563a9b90e545cdbf39527c5: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:22.817738 containerd[1899]: time="2026-03-02T12:50:22.817695973Z" level=info msg="CreateContainer within sandbox \"8dbcd91caca89281904690868489e21043c780aeda1e1d49a2ba3b9dd88bf6c2\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"1c856bf3d53ce0fd79e61b57cd6663c8ccce85378563a9b90e545cdbf39527c5\"" Mar 2 12:50:22.818583 containerd[1899]: time="2026-03-02T12:50:22.818552402Z" level=info msg="StartContainer for \"1c856bf3d53ce0fd79e61b57cd6663c8ccce85378563a9b90e545cdbf39527c5\"" Mar 2 12:50:22.819292 containerd[1899]: time="2026-03-02T12:50:22.819264066Z" level=info msg="connecting to shim 1c856bf3d53ce0fd79e61b57cd6663c8ccce85378563a9b90e545cdbf39527c5" address="unix:///run/containerd/s/52dd938832a1047e158113ed957b20895882146d45e92b5dc8f59e062af6886c" protocol=ttrpc version=3 Mar 2 12:50:22.840601 systemd[1]: Started cri-containerd-1c856bf3d53ce0fd79e61b57cd6663c8ccce85378563a9b90e545cdbf39527c5.scope - libcontainer container 1c856bf3d53ce0fd79e61b57cd6663c8ccce85378563a9b90e545cdbf39527c5. Mar 2 12:50:22.872820 containerd[1899]: time="2026-03-02T12:50:22.872716405Z" level=info msg="StartContainer for \"1c856bf3d53ce0fd79e61b57cd6663c8ccce85378563a9b90e545cdbf39527c5\" returns successfully" Mar 2 12:50:23.608286 kubelet[3475]: I0302 12:50:23.608089 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/whisker-cd5749c56-698wr" podStartSLOduration=10.768825658 podStartE2EDuration="20.608078654s" podCreationTimestamp="2026-03-02 12:50:03 +0000 UTC" firstStartedPulling="2026-03-02 12:50:07.495732802 +0000 UTC m=+48.198063541" lastFinishedPulling="2026-03-02 12:50:17.33498579 +0000 UTC m=+58.037316537" observedRunningTime="2026-03-02 12:50:19.62515034 +0000 UTC m=+60.327481079" watchObservedRunningTime="2026-03-02 12:50:23.608078654 +0000 UTC m=+64.310409393" Mar 2 12:50:23.643195 kubelet[3475]: I0302 12:50:23.643150 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6b7cf495fc-ftpl9" podStartSLOduration=36.921624624 podStartE2EDuration="44.643141319s" podCreationTimestamp="2026-03-02 12:49:39 +0000 UTC" firstStartedPulling="2026-03-02 12:50:15.035048721 +0000 UTC m=+55.737379460" lastFinishedPulling="2026-03-02 12:50:22.756565416 +0000 UTC m=+63.458896155" observedRunningTime="2026-03-02 12:50:23.608967179 +0000 UTC m=+64.311297918" watchObservedRunningTime="2026-03-02 12:50:23.643141319 +0000 UTC m=+64.345472066" Mar 2 12:50:25.358467 containerd[1899]: time="2026-03-02T12:50:25.357926067Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:25.364934 containerd[1899]: time="2026-03-02T12:50:25.364910565Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.3: active requests=0, bytes read=45512258" Mar 2 12:50:25.372203 containerd[1899]: time="2026-03-02T12:50:25.372179049Z" level=info msg="ImageCreate event name:\"sha256:6c1d6f109ccbdc040de9bade4e1d6f18ad2b7e93a2479f2ff827985a6b5c9653\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:25.381384 containerd[1899]: time="2026-03-02T12:50:25.381353678Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:c2def03be7412561bd678df17fcf2467cac990dbb42278dcfe193aa5a43128d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:25.382071 containerd[1899]: time="2026-03-02T12:50:25.381793372Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" with image id \"sha256:6c1d6f109ccbdc040de9bade4e1d6f18ad2b7e93a2479f2ff827985a6b5c9653\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:c2def03be7412561bd678df17fcf2467cac990dbb42278dcfe193aa5a43128d4\", size \"46909799\" in 2.624840071s" Mar 2 12:50:25.383201 containerd[1899]: time="2026-03-02T12:50:25.382982260Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" returns image reference \"sha256:6c1d6f109ccbdc040de9bade4e1d6f18ad2b7e93a2479f2ff827985a6b5c9653\"" Mar 2 12:50:25.384920 containerd[1899]: time="2026-03-02T12:50:25.384707102Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\"" Mar 2 12:50:25.392550 containerd[1899]: time="2026-03-02T12:50:25.392528037Z" level=info msg="CreateContainer within sandbox \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 2 12:50:25.415362 containerd[1899]: time="2026-03-02T12:50:25.415334723Z" level=info msg="Container 3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:25.417830 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount408789486.mount: Deactivated successfully. Mar 2 12:50:25.434863 containerd[1899]: time="2026-03-02T12:50:25.434835458Z" level=info msg="CreateContainer within sandbox \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\"" Mar 2 12:50:25.435363 containerd[1899]: time="2026-03-02T12:50:25.435308386Z" level=info msg="StartContainer for \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\"" Mar 2 12:50:25.437416 containerd[1899]: time="2026-03-02T12:50:25.437321197Z" level=info msg="connecting to shim 3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9" address="unix:///run/containerd/s/48219788bc27c5f50459db8160775c0ceede6cffe029185097fe9d92baf7dcd6" protocol=ttrpc version=3 Mar 2 12:50:25.456604 systemd[1]: Started cri-containerd-3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9.scope - libcontainer container 3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9. Mar 2 12:50:25.491791 containerd[1899]: time="2026-03-02T12:50:25.491730176Z" level=info msg="StartContainer for \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" returns successfully" Mar 2 12:50:25.726641 containerd[1899]: time="2026-03-02T12:50:25.726553999Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:25.730173 containerd[1899]: time="2026-03-02T12:50:25.730149007Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.3: active requests=0, bytes read=77" Mar 2 12:50:25.731811 containerd[1899]: time="2026-03-02T12:50:25.731740261Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" with image id \"sha256:6c1d6f109ccbdc040de9bade4e1d6f18ad2b7e93a2479f2ff827985a6b5c9653\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:c2def03be7412561bd678df17fcf2467cac990dbb42278dcfe193aa5a43128d4\", size \"46909799\" in 347.004509ms" Mar 2 12:50:25.731811 containerd[1899]: time="2026-03-02T12:50:25.731765710Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" returns image reference \"sha256:6c1d6f109ccbdc040de9bade4e1d6f18ad2b7e93a2479f2ff827985a6b5c9653\"" Mar 2 12:50:25.733157 containerd[1899]: time="2026-03-02T12:50:25.732867443Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.3\"" Mar 2 12:50:25.743719 containerd[1899]: time="2026-03-02T12:50:25.743698614Z" level=info msg="CreateContainer within sandbox \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 2 12:50:25.772797 containerd[1899]: time="2026-03-02T12:50:25.772769399Z" level=info msg="Container 1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:25.774653 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3141587063.mount: Deactivated successfully. Mar 2 12:50:25.794710 containerd[1899]: time="2026-03-02T12:50:25.794638221Z" level=info msg="CreateContainer within sandbox \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\"" Mar 2 12:50:25.795391 containerd[1899]: time="2026-03-02T12:50:25.795048259Z" level=info msg="StartContainer for \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\"" Mar 2 12:50:25.796114 containerd[1899]: time="2026-03-02T12:50:25.796074333Z" level=info msg="connecting to shim 1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8" address="unix:///run/containerd/s/6043b40311a0eac475e7477ae43eda7d8dca1c7b3475c4502f9a62f33f52c990" protocol=ttrpc version=3 Mar 2 12:50:25.812596 systemd[1]: Started cri-containerd-1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8.scope - libcontainer container 1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8. Mar 2 12:50:25.851404 containerd[1899]: time="2026-03-02T12:50:25.851381471Z" level=info msg="StartContainer for \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" returns successfully" Mar 2 12:50:26.662972 kubelet[3475]: I0302 12:50:26.662912 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-8487494cdb-p2qk6" podStartSLOduration=39.375642723 podStartE2EDuration="48.662896677s" podCreationTimestamp="2026-03-02 12:49:38 +0000 UTC" firstStartedPulling="2026-03-02 12:50:16.096822679 +0000 UTC m=+56.799153426" lastFinishedPulling="2026-03-02 12:50:25.384076633 +0000 UTC m=+66.086407380" observedRunningTime="2026-03-02 12:50:25.616854843 +0000 UTC m=+66.319185590" watchObservedRunningTime="2026-03-02 12:50:26.662896677 +0000 UTC m=+67.365227424" Mar 2 12:50:26.959692 kubelet[3475]: I0302 12:50:26.958924 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-8487494cdb-7zh2j" podStartSLOduration=41.621073716 podStartE2EDuration="48.958913499s" podCreationTimestamp="2026-03-02 12:49:38 +0000 UTC" firstStartedPulling="2026-03-02 12:50:18.3948089 +0000 UTC m=+59.097139639" lastFinishedPulling="2026-03-02 12:50:25.732648675 +0000 UTC m=+66.434979422" observedRunningTime="2026-03-02 12:50:26.663442631 +0000 UTC m=+67.365773378" watchObservedRunningTime="2026-03-02 12:50:26.958913499 +0000 UTC m=+67.661244238" Mar 2 12:50:28.513043 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount696902136.mount: Deactivated successfully. Mar 2 12:50:28.838518 containerd[1899]: time="2026-03-02T12:50:28.838196318Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:28.843126 containerd[1899]: time="2026-03-02T12:50:28.843097354Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.3: active requests=0, bytes read=51600693" Mar 2 12:50:28.846579 containerd[1899]: time="2026-03-02T12:50:28.846541072Z" level=info msg="ImageCreate event name:\"sha256:d40b2a23702c4c62ef242fb10a0dae8b80d5b5a0fd36ecec29e43b227f22611d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:28.851267 containerd[1899]: time="2026-03-02T12:50:28.851230974Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:e85ffa1d9468908b0bd44664de0d023da6669faefb3e1013b3a15b63dfa1f9a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:28.852230 containerd[1899]: time="2026-03-02T12:50:28.852139755Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.3\" with image id \"sha256:d40b2a23702c4c62ef242fb10a0dae8b80d5b5a0fd36ecec29e43b227f22611d\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:e85ffa1d9468908b0bd44664de0d023da6669faefb3e1013b3a15b63dfa1f9a9\", size \"51600539\" in 3.119249032s" Mar 2 12:50:28.852230 containerd[1899]: time="2026-03-02T12:50:28.852165644Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.3\" returns image reference \"sha256:d40b2a23702c4c62ef242fb10a0dae8b80d5b5a0fd36ecec29e43b227f22611d\"" Mar 2 12:50:28.853568 containerd[1899]: time="2026-03-02T12:50:28.853362426Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\"" Mar 2 12:50:28.860341 containerd[1899]: time="2026-03-02T12:50:28.860315121Z" level=info msg="CreateContainer within sandbox \"68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Mar 2 12:50:28.884173 containerd[1899]: time="2026-03-02T12:50:28.883647674Z" level=info msg="Container 828520f07b86aabe35da232d12488b84e4b1b60fea15194758217a4e102d7b14: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:28.901324 containerd[1899]: time="2026-03-02T12:50:28.901301679Z" level=info msg="CreateContainer within sandbox \"68f09ffe6fe507e4b0ad85b893f25e17ca511261ad465f043225e76201bbb2ed\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"828520f07b86aabe35da232d12488b84e4b1b60fea15194758217a4e102d7b14\"" Mar 2 12:50:28.902001 containerd[1899]: time="2026-03-02T12:50:28.901983396Z" level=info msg="StartContainer for \"828520f07b86aabe35da232d12488b84e4b1b60fea15194758217a4e102d7b14\"" Mar 2 12:50:28.903642 containerd[1899]: time="2026-03-02T12:50:28.903613504Z" level=info msg="connecting to shim 828520f07b86aabe35da232d12488b84e4b1b60fea15194758217a4e102d7b14" address="unix:///run/containerd/s/426942d99f1e521528ecfdaf82b5014db3bd9be5db92197c3393535aa40933d0" protocol=ttrpc version=3 Mar 2 12:50:28.938616 systemd[1]: Started cri-containerd-828520f07b86aabe35da232d12488b84e4b1b60fea15194758217a4e102d7b14.scope - libcontainer container 828520f07b86aabe35da232d12488b84e4b1b60fea15194758217a4e102d7b14. Mar 2 12:50:28.973391 containerd[1899]: time="2026-03-02T12:50:28.973361790Z" level=info msg="StartContainer for \"828520f07b86aabe35da232d12488b84e4b1b60fea15194758217a4e102d7b14\" returns successfully" Mar 2 12:50:29.240597 containerd[1899]: time="2026-03-02T12:50:29.240360092Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:29.243233 containerd[1899]: time="2026-03-02T12:50:29.243204471Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.3: active requests=0, bytes read=77" Mar 2 12:50:29.244437 containerd[1899]: time="2026-03-02T12:50:29.244408501Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" with image id \"sha256:6c1d6f109ccbdc040de9bade4e1d6f18ad2b7e93a2479f2ff827985a6b5c9653\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:c2def03be7412561bd678df17fcf2467cac990dbb42278dcfe193aa5a43128d4\", size \"46909799\" in 391.022778ms" Mar 2 12:50:29.244437 containerd[1899]: time="2026-03-02T12:50:29.244439014Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.3\" returns image reference \"sha256:6c1d6f109ccbdc040de9bade4e1d6f18ad2b7e93a2479f2ff827985a6b5c9653\"" Mar 2 12:50:29.245387 containerd[1899]: time="2026-03-02T12:50:29.245369508Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\"" Mar 2 12:50:29.252826 containerd[1899]: time="2026-03-02T12:50:29.252794977Z" level=info msg="CreateContainer within sandbox \"dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 2 12:50:29.283400 containerd[1899]: time="2026-03-02T12:50:29.283374507Z" level=info msg="Container a791d7a5a7ef418e0bad9e5f6fff5d23825553f28526c9ffc22e7781078a24aa: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:29.300736 containerd[1899]: time="2026-03-02T12:50:29.300707781Z" level=info msg="CreateContainer within sandbox \"dc16c6b49d5d6f4f1be87114efa67686ff80204ac0617e64bac7c13e270fafb0\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"a791d7a5a7ef418e0bad9e5f6fff5d23825553f28526c9ffc22e7781078a24aa\"" Mar 2 12:50:29.301844 containerd[1899]: time="2026-03-02T12:50:29.301811872Z" level=info msg="StartContainer for \"a791d7a5a7ef418e0bad9e5f6fff5d23825553f28526c9ffc22e7781078a24aa\"" Mar 2 12:50:29.302626 containerd[1899]: time="2026-03-02T12:50:29.302559592Z" level=info msg="connecting to shim a791d7a5a7ef418e0bad9e5f6fff5d23825553f28526c9ffc22e7781078a24aa" address="unix:///run/containerd/s/f793bff982e5b0e95978a8cedbdd3783d5985c138deb1b3178629b5cf231feb5" protocol=ttrpc version=3 Mar 2 12:50:29.319609 systemd[1]: Started cri-containerd-a791d7a5a7ef418e0bad9e5f6fff5d23825553f28526c9ffc22e7781078a24aa.scope - libcontainer container a791d7a5a7ef418e0bad9e5f6fff5d23825553f28526c9ffc22e7781078a24aa. Mar 2 12:50:29.363738 containerd[1899]: time="2026-03-02T12:50:29.363627768Z" level=info msg="StartContainer for \"a791d7a5a7ef418e0bad9e5f6fff5d23825553f28526c9ffc22e7781078a24aa\" returns successfully" Mar 2 12:50:29.682579 kubelet[3475]: I0302 12:50:29.682376 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-68bfcfb4bc-jjtz2" podStartSLOduration=40.0447678 podStartE2EDuration="50.682362819s" podCreationTimestamp="2026-03-02 12:49:39 +0000 UTC" firstStartedPulling="2026-03-02 12:50:18.607519241 +0000 UTC m=+59.309849988" lastFinishedPulling="2026-03-02 12:50:29.245114228 +0000 UTC m=+69.947445007" observedRunningTime="2026-03-02 12:50:29.677382268 +0000 UTC m=+70.379713015" watchObservedRunningTime="2026-03-02 12:50:29.682362819 +0000 UTC m=+70.384693566" Mar 2 12:50:30.665093 kubelet[3475]: I0302 12:50:30.664600 3475 prober_manager.go:356] "Failed to trigger a manual run" probe="Readiness" Mar 2 12:50:31.456344 containerd[1899]: time="2026-03-02T12:50:31.456300559Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:31.459626 containerd[1899]: time="2026-03-02T12:50:31.459484861Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3: active requests=0, bytes read=13755078" Mar 2 12:50:31.465288 containerd[1899]: time="2026-03-02T12:50:31.465260533Z" level=info msg="ImageCreate event name:\"sha256:c55251c1db32bbbf386d6ef9309a13d39443eef28f12c0883c2fd06bc5561b09\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:31.469936 containerd[1899]: time="2026-03-02T12:50:31.469914458Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:2bdced3111efc84af5b77534155b084a55a3f839010807e7e83e75faefc8cf33\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:50:31.470327 containerd[1899]: time="2026-03-02T12:50:31.470207699Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\" with image id \"sha256:c55251c1db32bbbf386d6ef9309a13d39443eef28f12c0883c2fd06bc5561b09\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:2bdced3111efc84af5b77534155b084a55a3f839010807e7e83e75faefc8cf33\", size \"15152555\" in 2.22481711s" Mar 2 12:50:31.470327 containerd[1899]: time="2026-03-02T12:50:31.470244965Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.3\" returns image reference \"sha256:c55251c1db32bbbf386d6ef9309a13d39443eef28f12c0883c2fd06bc5561b09\"" Mar 2 12:50:31.480003 containerd[1899]: time="2026-03-02T12:50:31.479777533Z" level=info msg="CreateContainer within sandbox \"c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 2 12:50:31.504434 containerd[1899]: time="2026-03-02T12:50:31.504409121Z" level=info msg="Container 8dfa85453da52a5a6c9d0c99b8a684e9c94ebd7e795594c070321e2942fc4b5f: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:31.532328 containerd[1899]: time="2026-03-02T12:50:31.532289724Z" level=info msg="CreateContainer within sandbox \"c808a28fb1c7b89b8bd43ecf82003eedcb1746d80af4b7cecafe761eb2d8b607\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"8dfa85453da52a5a6c9d0c99b8a684e9c94ebd7e795594c070321e2942fc4b5f\"" Mar 2 12:50:31.533496 containerd[1899]: time="2026-03-02T12:50:31.533319669Z" level=info msg="StartContainer for \"8dfa85453da52a5a6c9d0c99b8a684e9c94ebd7e795594c070321e2942fc4b5f\"" Mar 2 12:50:31.534384 containerd[1899]: time="2026-03-02T12:50:31.534358054Z" level=info msg="connecting to shim 8dfa85453da52a5a6c9d0c99b8a684e9c94ebd7e795594c070321e2942fc4b5f" address="unix:///run/containerd/s/b862d9a913fdd9bcc97e774726db3eba1b9272983c2647c18f376441208858ca" protocol=ttrpc version=3 Mar 2 12:50:31.555645 systemd[1]: Started cri-containerd-8dfa85453da52a5a6c9d0c99b8a684e9c94ebd7e795594c070321e2942fc4b5f.scope - libcontainer container 8dfa85453da52a5a6c9d0c99b8a684e9c94ebd7e795594c070321e2942fc4b5f. Mar 2 12:50:31.616033 containerd[1899]: time="2026-03-02T12:50:31.615994471Z" level=info msg="StartContainer for \"8dfa85453da52a5a6c9d0c99b8a684e9c94ebd7e795594c070321e2942fc4b5f\" returns successfully" Mar 2 12:50:31.681708 kubelet[3475]: I0302 12:50:31.681450 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/goldmane-7d7658d587-6g28g" podStartSLOduration=43.396054262 podStartE2EDuration="53.681440275s" podCreationTimestamp="2026-03-02 12:49:38 +0000 UTC" firstStartedPulling="2026-03-02 12:50:18.567341161 +0000 UTC m=+59.269671900" lastFinishedPulling="2026-03-02 12:50:28.852727174 +0000 UTC m=+69.555057913" observedRunningTime="2026-03-02 12:50:29.69439014 +0000 UTC m=+70.396720887" watchObservedRunningTime="2026-03-02 12:50:31.681440275 +0000 UTC m=+72.383771014" Mar 2 12:50:32.459269 kubelet[3475]: I0302 12:50:32.459091 3475 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 2 12:50:32.459269 kubelet[3475]: I0302 12:50:32.459125 3475 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 2 12:50:33.607126 kubelet[3475]: I0302 12:50:33.607035 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/csi-node-driver-wtzgs" podStartSLOduration=36.282156595000004 podStartE2EDuration="54.60702247s" podCreationTimestamp="2026-03-02 12:49:39 +0000 UTC" firstStartedPulling="2026-03-02 12:50:13.14619606 +0000 UTC m=+53.848526799" lastFinishedPulling="2026-03-02 12:50:31.471061935 +0000 UTC m=+72.173392674" observedRunningTime="2026-03-02 12:50:31.683099112 +0000 UTC m=+72.385429859" watchObservedRunningTime="2026-03-02 12:50:33.60702247 +0000 UTC m=+74.309353217" Mar 2 12:50:50.265996 kubelet[3475]: I0302 12:50:50.265817 3475 prober_manager.go:356] "Failed to trigger a manual run" probe="Readiness" Mar 2 12:50:50.316895 containerd[1899]: time="2026-03-02T12:50:50.316297464Z" level=info msg="StopContainer for \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" with timeout 30 (s)" Mar 2 12:50:50.320656 containerd[1899]: time="2026-03-02T12:50:50.320619041Z" level=info msg="Stop container \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" with signal terminated" Mar 2 12:50:50.344715 systemd[1]: cri-containerd-1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8.scope: Deactivated successfully. Mar 2 12:50:50.346838 containerd[1899]: time="2026-03-02T12:50:50.346626923Z" level=info msg="received container exit event container_id:\"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" id:\"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" pid:6050 exit_status:1 exited_at:{seconds:1772455850 nanos:345912132}" Mar 2 12:50:50.380689 systemd[1]: Created slice kubepods-besteffort-pod8cce4287_559b_4cb5_8f75_25bd267baeff.slice - libcontainer container kubepods-besteffort-pod8cce4287_559b_4cb5_8f75_25bd267baeff.slice. Mar 2 12:50:50.386721 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8-rootfs.mount: Deactivated successfully. Mar 2 12:50:50.449760 kubelet[3475]: I0302 12:50:50.449725 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h452r\" (UniqueName: \"kubernetes.io/projected/8cce4287-559b-4cb5-8f75-25bd267baeff-kube-api-access-h452r\") pod \"calico-apiserver-68bfcfb4bc-99cpk\" (UID: \"8cce4287-559b-4cb5-8f75-25bd267baeff\") " pod="calico-system/calico-apiserver-68bfcfb4bc-99cpk" Mar 2 12:50:50.449760 kubelet[3475]: I0302 12:50:50.449763 3475 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/8cce4287-559b-4cb5-8f75-25bd267baeff-calico-apiserver-certs\") pod \"calico-apiserver-68bfcfb4bc-99cpk\" (UID: \"8cce4287-559b-4cb5-8f75-25bd267baeff\") " pod="calico-system/calico-apiserver-68bfcfb4bc-99cpk" Mar 2 12:50:51.096517 containerd[1899]: time="2026-03-02T12:50:51.096367889Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68bfcfb4bc-99cpk,Uid:8cce4287-559b-4cb5-8f75-25bd267baeff,Namespace:calico-system,Attempt:0,}" Mar 2 12:50:51.124514 containerd[1899]: time="2026-03-02T12:50:51.124416967Z" level=info msg="StopContainer for \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" returns successfully" Mar 2 12:50:51.128907 containerd[1899]: time="2026-03-02T12:50:51.128875157Z" level=info msg="StopPodSandbox for \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\"" Mar 2 12:50:51.139649 containerd[1899]: time="2026-03-02T12:50:51.139623374Z" level=info msg="Container to stop \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:50:51.145232 systemd[1]: cri-containerd-ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51.scope: Deactivated successfully. Mar 2 12:50:51.147798 containerd[1899]: time="2026-03-02T12:50:51.147742110Z" level=info msg="received sandbox exit event container_id:\"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" id:\"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" exit_status:137 exited_at:{seconds:1772455851 nanos:147527431}" monitor_name=podsandbox Mar 2 12:50:51.179566 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51-rootfs.mount: Deactivated successfully. Mar 2 12:50:51.182889 containerd[1899]: time="2026-03-02T12:50:51.182844961Z" level=info msg="shim disconnected" id=ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51 namespace=k8s.io Mar 2 12:50:51.183032 containerd[1899]: time="2026-03-02T12:50:51.182867306Z" level=warning msg="cleaning up after shim disconnected" id=ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51 namespace=k8s.io Mar 2 12:50:51.183032 containerd[1899]: time="2026-03-02T12:50:51.183006758Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 12:50:51.211002 systemd-networkd[1485]: cali21fdbbad9a8: Link UP Mar 2 12:50:51.211991 systemd-networkd[1485]: cali21fdbbad9a8: Gained carrier Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.134 [INFO][6367] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0 calico-apiserver-68bfcfb4bc- calico-system 8cce4287-559b-4cb5-8f75-25bd267baeff 1202 0 2026-03-02 12:50:50 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:68bfcfb4bc projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4459.2.101-47783670b7 calico-apiserver-68bfcfb4bc-99cpk eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali21fdbbad9a8 [] [] }} ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.134 [INFO][6367] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.158 [INFO][6379] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" HandleID="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.164 [INFO][6379] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" HandleID="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fb3a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4459.2.101-47783670b7", "pod":"calico-apiserver-68bfcfb4bc-99cpk", "timestamp":"2026-03-02 12:50:51.158293385 +0000 UTC"}, Hostname:"ci-4459.2.101-47783670b7", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000375080)} Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.164 [INFO][6379] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.164 [INFO][6379] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.164 [INFO][6379] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4459.2.101-47783670b7' Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.166 [INFO][6379] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.169 [INFO][6379] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.174 [INFO][6379] ipam/ipam.go 526: Trying affinity for 192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.178 [INFO][6379] ipam/ipam.go 160: Attempting to load block cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.181 [INFO][6379] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.54.64/26 host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.181 [INFO][6379] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.54.64/26 handle="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.183 [INFO][6379] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9 Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.191 [INFO][6379] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.54.64/26 handle="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.205 [INFO][6379] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.54.74/26] block=192.168.54.64/26 handle="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.205 [INFO][6379] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.54.74/26] handle="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" host="ci-4459.2.101-47783670b7" Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.206 [INFO][6379] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:51.232996 containerd[1899]: 2026-03-02 12:50:51.206 [INFO][6379] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.54.74/26] IPv6=[] ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" HandleID="k8s-pod-network.c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" Mar 2 12:50:51.238339 containerd[1899]: 2026-03-02 12:50:51.208 [INFO][6367] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0", GenerateName:"calico-apiserver-68bfcfb4bc-", Namespace:"calico-system", SelfLink:"", UID:"8cce4287-559b-4cb5-8f75-25bd267baeff", ResourceVersion:"1202", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 50, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68bfcfb4bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"", Pod:"calico-apiserver-68bfcfb4bc-99cpk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.74/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali21fdbbad9a8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:51.238339 containerd[1899]: 2026-03-02 12:50:51.208 [INFO][6367] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.54.74/32] ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" Mar 2 12:50:51.238339 containerd[1899]: 2026-03-02 12:50:51.208 [INFO][6367] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali21fdbbad9a8 ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" Mar 2 12:50:51.238339 containerd[1899]: 2026-03-02 12:50:51.214 [INFO][6367] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" Mar 2 12:50:51.238339 containerd[1899]: 2026-03-02 12:50:51.215 [INFO][6367] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0", GenerateName:"calico-apiserver-68bfcfb4bc-", Namespace:"calico-system", SelfLink:"", UID:"8cce4287-559b-4cb5-8f75-25bd267baeff", ResourceVersion:"1202", Generation:0, CreationTimestamp:time.Date(2026, time.March, 2, 12, 50, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"68bfcfb4bc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4459.2.101-47783670b7", ContainerID:"c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9", Pod:"calico-apiserver-68bfcfb4bc-99cpk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.54.74/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali21fdbbad9a8", MAC:"26:33:1f:7d:15:34", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 2 12:50:51.238339 containerd[1899]: 2026-03-02 12:50:51.228 [INFO][6367] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" Namespace="calico-system" Pod="calico-apiserver-68bfcfb4bc-99cpk" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--68bfcfb4bc--99cpk-eth0" Mar 2 12:50:51.252359 containerd[1899]: time="2026-03-02T12:50:51.250836564Z" level=info msg="received sandbox container exit event sandbox_id:\"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" exit_status:137 exited_at:{seconds:1772455851 nanos:147527431}" monitor_name=criService Mar 2 12:50:51.252953 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51-shm.mount: Deactivated successfully. Mar 2 12:50:51.286373 containerd[1899]: time="2026-03-02T12:50:51.286339084Z" level=info msg="connecting to shim c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9" address="unix:///run/containerd/s/6171ec7059cc7528104be8f9d73dc705daf83b6897d13f799189005465b6b727" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:50:51.294289 systemd-networkd[1485]: cali3b23f72654d: Link DOWN Mar 2 12:50:51.294297 systemd-networkd[1485]: cali3b23f72654d: Lost carrier Mar 2 12:50:51.317356 systemd[1]: Started cri-containerd-c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9.scope - libcontainer container c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9. Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.289 [INFO][6442] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.290 [INFO][6442] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" iface="eth0" netns="/var/run/netns/cni-4300c3a6-2bb2-6d67-4073-7bb399fdabb8" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.292 [INFO][6442] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" iface="eth0" netns="/var/run/netns/cni-4300c3a6-2bb2-6d67-4073-7bb399fdabb8" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.299 [INFO][6442] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" after=8.571392ms iface="eth0" netns="/var/run/netns/cni-4300c3a6-2bb2-6d67-4073-7bb399fdabb8" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.299 [INFO][6442] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.299 [INFO][6442] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.328 [INFO][6478] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.329 [INFO][6478] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.329 [INFO][6478] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.384 [INFO][6478] ipam/ipam_plugin.go 516: Released address using handleID ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.384 [INFO][6478] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.389 [INFO][6478] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:51.396164 containerd[1899]: 2026-03-02 12:50:51.393 [INFO][6442] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:50:51.397966 systemd[1]: run-netns-cni\x2d4300c3a6\x2d2bb2\x2d6d67\x2d4073\x2d7bb399fdabb8.mount: Deactivated successfully. Mar 2 12:50:51.399526 containerd[1899]: time="2026-03-02T12:50:51.399297085Z" level=info msg="TearDown network for sandbox \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" successfully" Mar 2 12:50:51.399526 containerd[1899]: time="2026-03-02T12:50:51.399321045Z" level=info msg="StopPodSandbox for \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" returns successfully" Mar 2 12:50:51.402393 containerd[1899]: time="2026-03-02T12:50:51.402365460Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-68bfcfb4bc-99cpk,Uid:8cce4287-559b-4cb5-8f75-25bd267baeff,Namespace:calico-system,Attempt:0,} returns sandbox id \"c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9\"" Mar 2 12:50:51.414908 containerd[1899]: time="2026-03-02T12:50:51.414861279Z" level=info msg="CreateContainer within sandbox \"c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 2 12:50:51.445277 containerd[1899]: time="2026-03-02T12:50:51.444769299Z" level=info msg="Container 82980b4bd2ef9c9a8535949678847a37ba42e9b9bdd6832da489c6b6b0395d2c: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:50:51.472473 containerd[1899]: time="2026-03-02T12:50:51.472427508Z" level=info msg="CreateContainer within sandbox \"c61cf88996f986cb4095c338a426637c4f01fba51a84d3a18306d66bdba4cfc9\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"82980b4bd2ef9c9a8535949678847a37ba42e9b9bdd6832da489c6b6b0395d2c\"" Mar 2 12:50:51.472904 containerd[1899]: time="2026-03-02T12:50:51.472882235Z" level=info msg="StartContainer for \"82980b4bd2ef9c9a8535949678847a37ba42e9b9bdd6832da489c6b6b0395d2c\"" Mar 2 12:50:51.474824 containerd[1899]: time="2026-03-02T12:50:51.474797428Z" level=info msg="connecting to shim 82980b4bd2ef9c9a8535949678847a37ba42e9b9bdd6832da489c6b6b0395d2c" address="unix:///run/containerd/s/6171ec7059cc7528104be8f9d73dc705daf83b6897d13f799189005465b6b727" protocol=ttrpc version=3 Mar 2 12:50:51.530627 systemd[1]: Started cri-containerd-82980b4bd2ef9c9a8535949678847a37ba42e9b9bdd6832da489c6b6b0395d2c.scope - libcontainer container 82980b4bd2ef9c9a8535949678847a37ba42e9b9bdd6832da489c6b6b0395d2c. Mar 2 12:50:51.557920 kubelet[3475]: I0302 12:50:51.557839 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/secret/56b6d31a-681f-4836-a653-77f50d324a86-calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/56b6d31a-681f-4836-a653-77f50d324a86-calico-apiserver-certs\") pod \"56b6d31a-681f-4836-a653-77f50d324a86\" (UID: \"56b6d31a-681f-4836-a653-77f50d324a86\") " Mar 2 12:50:51.557920 kubelet[3475]: I0302 12:50:51.557875 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/56b6d31a-681f-4836-a653-77f50d324a86-kube-api-access-dzv8c\" (UniqueName: \"kubernetes.io/projected/56b6d31a-681f-4836-a653-77f50d324a86-kube-api-access-dzv8c\") pod \"56b6d31a-681f-4836-a653-77f50d324a86\" (UID: \"56b6d31a-681f-4836-a653-77f50d324a86\") " Mar 2 12:50:51.561335 kubelet[3475]: I0302 12:50:51.561277 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b6d31a-681f-4836-a653-77f50d324a86-calico-apiserver-certs" pod "56b6d31a-681f-4836-a653-77f50d324a86" (UID: "56b6d31a-681f-4836-a653-77f50d324a86"). InnerVolumeSpecName "calico-apiserver-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 2 12:50:51.561335 kubelet[3475]: I0302 12:50:51.561318 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56b6d31a-681f-4836-a653-77f50d324a86-kube-api-access-dzv8c" pod "56b6d31a-681f-4836-a653-77f50d324a86" (UID: "56b6d31a-681f-4836-a653-77f50d324a86"). InnerVolumeSpecName "kube-api-access-dzv8c". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 12:50:51.564080 containerd[1899]: time="2026-03-02T12:50:51.564001215Z" level=info msg="StartContainer for \"82980b4bd2ef9c9a8535949678847a37ba42e9b9bdd6832da489c6b6b0395d2c\" returns successfully" Mar 2 12:50:51.658596 kubelet[3475]: I0302 12:50:51.658473 3475 reconciler_common.go:299] "Volume detached for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/56b6d31a-681f-4836-a653-77f50d324a86-calico-apiserver-certs\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:51.658596 kubelet[3475]: I0302 12:50:51.658513 3475 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dzv8c\" (UniqueName: \"kubernetes.io/projected/56b6d31a-681f-4836-a653-77f50d324a86-kube-api-access-dzv8c\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:51.721701 kubelet[3475]: I0302 12:50:51.721584 3475 scope.go:122] "RemoveContainer" containerID="1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8" Mar 2 12:50:51.725064 containerd[1899]: time="2026-03-02T12:50:51.725038918Z" level=info msg="RemoveContainer for \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\"" Mar 2 12:50:51.726726 systemd[1]: Removed slice kubepods-besteffort-pod56b6d31a_681f_4836_a653_77f50d324a86.slice - libcontainer container kubepods-besteffort-pod56b6d31a_681f_4836_a653_77f50d324a86.slice. Mar 2 12:50:51.749848 containerd[1899]: time="2026-03-02T12:50:51.749772636Z" level=info msg="RemoveContainer for \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" returns successfully" Mar 2 12:50:51.750313 kubelet[3475]: I0302 12:50:51.750295 3475 scope.go:122] "RemoveContainer" containerID="1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8" Mar 2 12:50:51.751389 containerd[1899]: time="2026-03-02T12:50:51.750819927Z" level=error msg="ContainerStatus for \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\": not found" Mar 2 12:50:51.751586 kubelet[3475]: E0302 12:50:51.751569 3475 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\": not found" containerID="1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8" Mar 2 12:50:51.751899 kubelet[3475]: I0302 12:50:51.751863 3475 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8"} err="failed to get container status \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\": rpc error: code = NotFound desc = an error occurred when try to find container \"1c8c469458ce3c9120b7b58bc7eb593fabec1622cbe5d83e2e06d326f315afa8\": not found" Mar 2 12:50:51.753644 kubelet[3475]: I0302 12:50:51.753602 3475 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="calico-system/calico-apiserver-68bfcfb4bc-99cpk" podStartSLOduration=1.753486449 podStartE2EDuration="1.753486449s" podCreationTimestamp="2026-03-02 12:50:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:50:51.735107216 +0000 UTC m=+92.437438067" watchObservedRunningTime="2026-03-02 12:50:51.753486449 +0000 UTC m=+92.455817188" Mar 2 12:50:52.387149 systemd[1]: var-lib-kubelet-pods-56b6d31a\x2d681f\x2d4836\x2da653\x2d77f50d324a86-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ddzv8c.mount: Deactivated successfully. Mar 2 12:50:52.387466 systemd[1]: var-lib-kubelet-pods-56b6d31a\x2d681f\x2d4836\x2da653\x2d77f50d324a86-volumes-kubernetes.io\x7esecret-calico\x2dapiserver\x2dcerts.mount: Deactivated successfully. Mar 2 12:50:52.391149 containerd[1899]: time="2026-03-02T12:50:52.390661410Z" level=info msg="StopContainer for \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" with timeout 30 (s)" Mar 2 12:50:52.391574 containerd[1899]: time="2026-03-02T12:50:52.391537271Z" level=info msg="Stop container \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" with signal terminated" Mar 2 12:50:52.413421 systemd[1]: cri-containerd-3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9.scope: Deactivated successfully. Mar 2 12:50:52.416464 containerd[1899]: time="2026-03-02T12:50:52.416247285Z" level=info msg="received container exit event container_id:\"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" id:\"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" pid:6011 exit_status:1 exited_at:{seconds:1772455852 nanos:415763189}" Mar 2 12:50:52.435082 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9-rootfs.mount: Deactivated successfully. Mar 2 12:50:52.455473 containerd[1899]: time="2026-03-02T12:50:52.455386671Z" level=info msg="StopContainer for \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" returns successfully" Mar 2 12:50:52.456038 containerd[1899]: time="2026-03-02T12:50:52.456010012Z" level=info msg="StopPodSandbox for \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\"" Mar 2 12:50:52.456151 containerd[1899]: time="2026-03-02T12:50:52.456134272Z" level=info msg="Container to stop \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:50:52.460895 systemd[1]: cri-containerd-bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423.scope: Deactivated successfully. Mar 2 12:50:52.462904 containerd[1899]: time="2026-03-02T12:50:52.462882691Z" level=info msg="received sandbox exit event container_id:\"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" id:\"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" exit_status:137 exited_at:{seconds:1772455852 nanos:462645043}" monitor_name=podsandbox Mar 2 12:50:52.478016 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423-rootfs.mount: Deactivated successfully. Mar 2 12:50:52.478328 containerd[1899]: time="2026-03-02T12:50:52.478275872Z" level=info msg="shim disconnected" id=bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423 namespace=k8s.io Mar 2 12:50:52.478422 containerd[1899]: time="2026-03-02T12:50:52.478294464Z" level=warning msg="cleaning up after shim disconnected" id=bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423 namespace=k8s.io Mar 2 12:50:52.478582 containerd[1899]: time="2026-03-02T12:50:52.478459422Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 12:50:52.501691 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423-shm.mount: Deactivated successfully. Mar 2 12:50:52.506120 containerd[1899]: time="2026-03-02T12:50:52.506085885Z" level=info msg="received sandbox container exit event sandbox_id:\"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" exit_status:137 exited_at:{seconds:1772455852 nanos:462645043}" monitor_name=criService Mar 2 12:50:52.541990 systemd-networkd[1485]: cali01cb78abbad: Link DOWN Mar 2 12:50:52.541997 systemd-networkd[1485]: cali01cb78abbad: Lost carrier Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.540 [INFO][6621] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.541 [INFO][6621] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" iface="eth0" netns="/var/run/netns/cni-aac2eb77-3ed0-38eb-02e8-d4e583aaa940" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.541 [INFO][6621] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" iface="eth0" netns="/var/run/netns/cni-aac2eb77-3ed0-38eb-02e8-d4e583aaa940" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.546 [INFO][6621] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" after=5.636718ms iface="eth0" netns="/var/run/netns/cni-aac2eb77-3ed0-38eb-02e8-d4e583aaa940" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.546 [INFO][6621] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.546 [INFO][6621] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.565 [INFO][6631] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.565 [INFO][6631] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.565 [INFO][6631] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.598 [INFO][6631] ipam/ipam_plugin.go 516: Released address using handleID ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.598 [INFO][6631] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.599 [INFO][6631] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:50:52.601746 containerd[1899]: 2026-03-02 12:50:52.600 [INFO][6621] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:50:52.603530 containerd[1899]: time="2026-03-02T12:50:52.602194696Z" level=info msg="TearDown network for sandbox \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" successfully" Mar 2 12:50:52.603530 containerd[1899]: time="2026-03-02T12:50:52.602218417Z" level=info msg="StopPodSandbox for \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" returns successfully" Mar 2 12:50:52.603904 systemd[1]: run-netns-cni\x2daac2eb77\x2d3ed0\x2d38eb\x2d02e8\x2dd4e583aaa940.mount: Deactivated successfully. Mar 2 12:50:52.668282 kubelet[3475]: I0302 12:50:52.667708 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/9d4f30bf-d266-4607-b976-3eb09820fbc8-kube-api-access-spjwv\" (UniqueName: \"kubernetes.io/projected/9d4f30bf-d266-4607-b976-3eb09820fbc8-kube-api-access-spjwv\") pod \"9d4f30bf-d266-4607-b976-3eb09820fbc8\" (UID: \"9d4f30bf-d266-4607-b976-3eb09820fbc8\") " Mar 2 12:50:52.668282 kubelet[3475]: I0302 12:50:52.668230 3475 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/secret/9d4f30bf-d266-4607-b976-3eb09820fbc8-calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/9d4f30bf-d266-4607-b976-3eb09820fbc8-calico-apiserver-certs\") pod \"9d4f30bf-d266-4607-b976-3eb09820fbc8\" (UID: \"9d4f30bf-d266-4607-b976-3eb09820fbc8\") " Mar 2 12:50:52.670514 kubelet[3475]: I0302 12:50:52.670172 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4f30bf-d266-4607-b976-3eb09820fbc8-kube-api-access-spjwv" pod "9d4f30bf-d266-4607-b976-3eb09820fbc8" (UID: "9d4f30bf-d266-4607-b976-3eb09820fbc8"). InnerVolumeSpecName "kube-api-access-spjwv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 12:50:52.672646 kubelet[3475]: I0302 12:50:52.672610 3475 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4f30bf-d266-4607-b976-3eb09820fbc8-calico-apiserver-certs" pod "9d4f30bf-d266-4607-b976-3eb09820fbc8" (UID: "9d4f30bf-d266-4607-b976-3eb09820fbc8"). InnerVolumeSpecName "calico-apiserver-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 2 12:50:52.725645 kubelet[3475]: I0302 12:50:52.724912 3475 scope.go:122] "RemoveContainer" containerID="3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9" Mar 2 12:50:52.728673 containerd[1899]: time="2026-03-02T12:50:52.728649526Z" level=info msg="RemoveContainer for \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\"" Mar 2 12:50:52.731458 systemd[1]: Removed slice kubepods-besteffort-pod9d4f30bf_d266_4607_b976_3eb09820fbc8.slice - libcontainer container kubepods-besteffort-pod9d4f30bf_d266_4607_b976_3eb09820fbc8.slice. Mar 2 12:50:52.739549 containerd[1899]: time="2026-03-02T12:50:52.739524027Z" level=info msg="RemoveContainer for \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" returns successfully" Mar 2 12:50:52.739763 kubelet[3475]: I0302 12:50:52.739741 3475 scope.go:122] "RemoveContainer" containerID="3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9" Mar 2 12:50:52.739970 containerd[1899]: time="2026-03-02T12:50:52.739944369Z" level=error msg="ContainerStatus for \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\": not found" Mar 2 12:50:52.740118 kubelet[3475]: E0302 12:50:52.740085 3475 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\": not found" containerID="3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9" Mar 2 12:50:52.740192 kubelet[3475]: I0302 12:50:52.740108 3475 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9"} err="failed to get container status \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\": rpc error: code = NotFound desc = an error occurred when try to find container \"3787c058a143c131dc662c01c793b6e222dfec64bd43a1b51607a254a34996c9\": not found" Mar 2 12:50:52.769908 kubelet[3475]: I0302 12:50:52.769860 3475 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-spjwv\" (UniqueName: \"kubernetes.io/projected/9d4f30bf-d266-4607-b976-3eb09820fbc8-kube-api-access-spjwv\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:52.770017 kubelet[3475]: I0302 12:50:52.769999 3475 reconciler_common.go:299] "Volume detached for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/9d4f30bf-d266-4607-b976-3eb09820fbc8-calico-apiserver-certs\") on node \"ci-4459.2.101-47783670b7\" DevicePath \"\"" Mar 2 12:50:53.042620 systemd-networkd[1485]: cali21fdbbad9a8: Gained IPv6LL Mar 2 12:50:53.383954 kubelet[3475]: I0302 12:50:53.383524 3475 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="56b6d31a-681f-4836-a653-77f50d324a86" path="/var/lib/kubelet/pods/56b6d31a-681f-4836-a653-77f50d324a86/volumes" Mar 2 12:50:53.383954 kubelet[3475]: I0302 12:50:53.383796 3475 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="9d4f30bf-d266-4607-b976-3eb09820fbc8" path="/var/lib/kubelet/pods/9d4f30bf-d266-4607-b976-3eb09820fbc8/volumes" Mar 2 12:50:53.387927 systemd[1]: var-lib-kubelet-pods-9d4f30bf\x2dd266\x2d4607\x2db976\x2d3eb09820fbc8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dspjwv.mount: Deactivated successfully. Mar 2 12:50:53.388020 systemd[1]: var-lib-kubelet-pods-9d4f30bf\x2dd266\x2d4607\x2db976\x2d3eb09820fbc8-volumes-kubernetes.io\x7esecret-calico\x2dapiserver\x2dcerts.mount: Deactivated successfully. Mar 2 12:51:02.533687 systemd[1]: Started sshd@7-10.200.20.38:22-10.200.16.10:50394.service - OpenSSH per-connection server daemon (10.200.16.10:50394). Mar 2 12:51:02.954748 sshd[6726]: Accepted publickey for core from 10.200.16.10 port 50394 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:02.956466 sshd-session[6726]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:02.960182 systemd-logind[1870]: New session 10 of user core. Mar 2 12:51:02.968620 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 2 12:51:03.237675 sshd[6735]: Connection closed by 10.200.16.10 port 50394 Mar 2 12:51:03.238294 sshd-session[6726]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:03.241288 systemd[1]: sshd@7-10.200.20.38:22-10.200.16.10:50394.service: Deactivated successfully. Mar 2 12:51:03.242849 systemd[1]: session-10.scope: Deactivated successfully. Mar 2 12:51:03.244961 systemd-logind[1870]: Session 10 logged out. Waiting for processes to exit. Mar 2 12:51:03.246894 systemd-logind[1870]: Removed session 10. Mar 2 12:51:08.327418 systemd[1]: Started sshd@8-10.200.20.38:22-10.200.16.10:50400.service - OpenSSH per-connection server daemon (10.200.16.10:50400). Mar 2 12:51:08.744266 sshd[6792]: Accepted publickey for core from 10.200.16.10 port 50400 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:08.745010 sshd-session[6792]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:08.748832 systemd-logind[1870]: New session 11 of user core. Mar 2 12:51:08.751615 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 2 12:51:09.017604 sshd[6795]: Connection closed by 10.200.16.10 port 50400 Mar 2 12:51:09.017804 sshd-session[6792]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:09.022173 systemd-logind[1870]: Session 11 logged out. Waiting for processes to exit. Mar 2 12:51:09.023881 systemd[1]: sshd@8-10.200.20.38:22-10.200.16.10:50400.service: Deactivated successfully. Mar 2 12:51:09.025324 systemd[1]: session-11.scope: Deactivated successfully. Mar 2 12:51:09.026904 systemd-logind[1870]: Removed session 11. Mar 2 12:51:14.110192 systemd[1]: Started sshd@9-10.200.20.38:22-10.200.16.10:41282.service - OpenSSH per-connection server daemon (10.200.16.10:41282). Mar 2 12:51:14.528454 sshd[6815]: Accepted publickey for core from 10.200.16.10 port 41282 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:14.529560 sshd-session[6815]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:14.532839 systemd-logind[1870]: New session 12 of user core. Mar 2 12:51:14.537792 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 2 12:51:14.803395 sshd[6818]: Connection closed by 10.200.16.10 port 41282 Mar 2 12:51:14.803625 sshd-session[6815]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:14.806939 systemd[1]: sshd@9-10.200.20.38:22-10.200.16.10:41282.service: Deactivated successfully. Mar 2 12:51:14.808236 systemd[1]: session-12.scope: Deactivated successfully. Mar 2 12:51:14.809300 systemd-logind[1870]: Session 12 logged out. Waiting for processes to exit. Mar 2 12:51:14.810349 systemd-logind[1870]: Removed session 12. Mar 2 12:51:19.366759 containerd[1899]: time="2026-03-02T12:51:19.366694692Z" level=info msg="StopPodSandbox for \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\"" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.398 [WARNING][6838] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.398 [INFO][6838] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.398 [INFO][6838] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" iface="eth0" netns="" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.398 [INFO][6838] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.398 [INFO][6838] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.415 [INFO][6849] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.415 [INFO][6849] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.415 [INFO][6849] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.430 [WARNING][6849] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.430 [INFO][6849] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.434 [INFO][6849] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:51:19.439524 containerd[1899]: 2026-03-02 12:51:19.437 [INFO][6838] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.439524 containerd[1899]: time="2026-03-02T12:51:19.439437284Z" level=info msg="TearDown network for sandbox \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" successfully" Mar 2 12:51:19.439524 containerd[1899]: time="2026-03-02T12:51:19.439454836Z" level=info msg="StopPodSandbox for \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" returns successfully" Mar 2 12:51:19.440536 containerd[1899]: time="2026-03-02T12:51:19.440512632Z" level=info msg="RemovePodSandbox for \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\"" Mar 2 12:51:19.440590 containerd[1899]: time="2026-03-02T12:51:19.440546041Z" level=info msg="Forcibly stopping sandbox \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\"" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.468 [WARNING][6864] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.468 [INFO][6864] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.468 [INFO][6864] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" iface="eth0" netns="" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.468 [INFO][6864] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.468 [INFO][6864] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.481 [INFO][6871] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.481 [INFO][6871] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.481 [INFO][6871] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.486 [WARNING][6871] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.486 [INFO][6871] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" HandleID="k8s-pod-network.ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--7zh2j-eth0" Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.487 [INFO][6871] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:51:19.489811 containerd[1899]: 2026-03-02 12:51:19.488 [INFO][6864] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51" Mar 2 12:51:19.490059 containerd[1899]: time="2026-03-02T12:51:19.489847519Z" level=info msg="TearDown network for sandbox \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" successfully" Mar 2 12:51:19.491126 containerd[1899]: time="2026-03-02T12:51:19.491105009Z" level=info msg="Ensure that sandbox ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51 in task-service has been cleanup successfully" Mar 2 12:51:19.506299 containerd[1899]: time="2026-03-02T12:51:19.506273861Z" level=info msg="RemovePodSandbox \"ff354798abd1ea276bb375212ca00ff2027d6caf510cf1f5d89e8d7768c6eb51\" returns successfully" Mar 2 12:51:19.506672 containerd[1899]: time="2026-03-02T12:51:19.506644706Z" level=info msg="StopPodSandbox for \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\"" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.530 [WARNING][6885] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.530 [INFO][6885] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.530 [INFO][6885] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" iface="eth0" netns="" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.530 [INFO][6885] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.530 [INFO][6885] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.543 [INFO][6892] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.543 [INFO][6892] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.543 [INFO][6892] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.548 [WARNING][6892] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.548 [INFO][6892] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.549 [INFO][6892] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:51:19.551373 containerd[1899]: 2026-03-02 12:51:19.550 [INFO][6885] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.551756 containerd[1899]: time="2026-03-02T12:51:19.551660952Z" level=info msg="TearDown network for sandbox \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" successfully" Mar 2 12:51:19.551756 containerd[1899]: time="2026-03-02T12:51:19.551682648Z" level=info msg="StopPodSandbox for \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" returns successfully" Mar 2 12:51:19.552208 containerd[1899]: time="2026-03-02T12:51:19.551980538Z" level=info msg="RemovePodSandbox for \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\"" Mar 2 12:51:19.552208 containerd[1899]: time="2026-03-02T12:51:19.552000467Z" level=info msg="Forcibly stopping sandbox \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\"" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.574 [WARNING][6906] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" WorkloadEndpoint="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.575 [INFO][6906] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.575 [INFO][6906] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" iface="eth0" netns="" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.575 [INFO][6906] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.575 [INFO][6906] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.588 [INFO][6913] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.588 [INFO][6913] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.588 [INFO][6913] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.592 [WARNING][6913] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.592 [INFO][6913] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" HandleID="k8s-pod-network.bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Workload="ci--4459.2.101--47783670b7-k8s-calico--apiserver--8487494cdb--p2qk6-eth0" Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.593 [INFO][6913] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 2 12:51:19.596224 containerd[1899]: 2026-03-02 12:51:19.595 [INFO][6906] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423" Mar 2 12:51:19.596630 containerd[1899]: time="2026-03-02T12:51:19.596405692Z" level=info msg="TearDown network for sandbox \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" successfully" Mar 2 12:51:19.597880 containerd[1899]: time="2026-03-02T12:51:19.597820652Z" level=info msg="Ensure that sandbox bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423 in task-service has been cleanup successfully" Mar 2 12:51:19.608964 containerd[1899]: time="2026-03-02T12:51:19.608937697Z" level=info msg="RemovePodSandbox \"bb8ca218258c6f65d44be955bb5582bcf01196a8d829abf9f250d181753cb423\" returns successfully" Mar 2 12:51:19.890983 systemd[1]: Started sshd@10-10.200.20.38:22-10.200.16.10:41292.service - OpenSSH per-connection server daemon (10.200.16.10:41292). Mar 2 12:51:20.311969 sshd[6920]: Accepted publickey for core from 10.200.16.10 port 41292 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:20.312985 sshd-session[6920]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:20.316560 systemd-logind[1870]: New session 13 of user core. Mar 2 12:51:20.322607 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 2 12:51:20.586880 sshd[6923]: Connection closed by 10.200.16.10 port 41292 Mar 2 12:51:20.587280 sshd-session[6920]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:20.590096 systemd[1]: sshd@10-10.200.20.38:22-10.200.16.10:41292.service: Deactivated successfully. Mar 2 12:51:20.591962 systemd[1]: session-13.scope: Deactivated successfully. Mar 2 12:51:20.592730 systemd-logind[1870]: Session 13 logged out. Waiting for processes to exit. Mar 2 12:51:20.593812 systemd-logind[1870]: Removed session 13. Mar 2 12:51:20.677878 systemd[1]: Started sshd@11-10.200.20.38:22-10.200.16.10:48704.service - OpenSSH per-connection server daemon (10.200.16.10:48704). Mar 2 12:51:21.093207 sshd[6935]: Accepted publickey for core from 10.200.16.10 port 48704 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:21.094131 sshd-session[6935]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:21.097645 systemd-logind[1870]: New session 14 of user core. Mar 2 12:51:21.101611 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 2 12:51:21.396160 sshd[6938]: Connection closed by 10.200.16.10 port 48704 Mar 2 12:51:21.397668 sshd-session[6935]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:21.402058 systemd[1]: sshd@11-10.200.20.38:22-10.200.16.10:48704.service: Deactivated successfully. Mar 2 12:51:21.403560 systemd[1]: session-14.scope: Deactivated successfully. Mar 2 12:51:21.406034 systemd-logind[1870]: Session 14 logged out. Waiting for processes to exit. Mar 2 12:51:21.408204 systemd-logind[1870]: Removed session 14. Mar 2 12:51:21.481686 systemd[1]: Started sshd@12-10.200.20.38:22-10.200.16.10:48720.service - OpenSSH per-connection server daemon (10.200.16.10:48720). Mar 2 12:51:21.896003 sshd[6962]: Accepted publickey for core from 10.200.16.10 port 48720 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:21.897159 sshd-session[6962]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:21.902969 systemd-logind[1870]: New session 15 of user core. Mar 2 12:51:21.905628 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 2 12:51:22.169767 sshd[6968]: Connection closed by 10.200.16.10 port 48720 Mar 2 12:51:22.170039 sshd-session[6962]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:22.173587 systemd[1]: sshd@12-10.200.20.38:22-10.200.16.10:48720.service: Deactivated successfully. Mar 2 12:51:22.175877 systemd[1]: session-15.scope: Deactivated successfully. Mar 2 12:51:22.178624 systemd-logind[1870]: Session 15 logged out. Waiting for processes to exit. Mar 2 12:51:22.180244 systemd-logind[1870]: Removed session 15. Mar 2 12:51:27.263216 systemd[1]: Started sshd@13-10.200.20.38:22-10.200.16.10:48728.service - OpenSSH per-connection server daemon (10.200.16.10:48728). Mar 2 12:51:27.672381 sshd[7016]: Accepted publickey for core from 10.200.16.10 port 48728 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:27.673373 sshd-session[7016]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:27.677531 systemd-logind[1870]: New session 16 of user core. Mar 2 12:51:27.683617 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 2 12:51:27.946555 sshd[7021]: Connection closed by 10.200.16.10 port 48728 Mar 2 12:51:27.946399 sshd-session[7016]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:27.950165 systemd[1]: sshd@13-10.200.20.38:22-10.200.16.10:48728.service: Deactivated successfully. Mar 2 12:51:27.952237 systemd[1]: session-16.scope: Deactivated successfully. Mar 2 12:51:27.953826 systemd-logind[1870]: Session 16 logged out. Waiting for processes to exit. Mar 2 12:51:27.955371 systemd-logind[1870]: Removed session 16. Mar 2 12:51:28.038332 systemd[1]: Started sshd@14-10.200.20.38:22-10.200.16.10:48744.service - OpenSSH per-connection server daemon (10.200.16.10:48744). Mar 2 12:51:28.452884 sshd[7033]: Accepted publickey for core from 10.200.16.10 port 48744 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:28.453590 sshd-session[7033]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:28.457160 systemd-logind[1870]: New session 17 of user core. Mar 2 12:51:28.468621 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 2 12:51:28.879212 sshd[7036]: Connection closed by 10.200.16.10 port 48744 Mar 2 12:51:28.895429 sshd-session[7033]: pam_unix(sshd:session): session closed for user core Mar 2 12:51:28.901087 systemd[1]: sshd@14-10.200.20.38:22-10.200.16.10:48744.service: Deactivated successfully. Mar 2 12:51:28.903402 systemd[1]: session-17.scope: Deactivated successfully. Mar 2 12:51:28.905929 systemd-logind[1870]: Session 17 logged out. Waiting for processes to exit. Mar 2 12:51:28.907942 systemd-logind[1870]: Removed session 17. Mar 2 12:51:28.970691 systemd[1]: Started sshd@15-10.200.20.38:22-10.200.16.10:48750.service - OpenSSH per-connection server daemon (10.200.16.10:48750). Mar 2 12:51:29.390184 sshd[7046]: Accepted publickey for core from 10.200.16.10 port 48750 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:51:29.391349 sshd-session[7046]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:51:29.398240 systemd-logind[1870]: New session 18 of user core. Mar 2 12:51:29.401719 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 2 12:52:02.233328 sshd[7049]: Connection closed by 10.200.16.10 port 48750 Mar 2 12:52:02.234085 sshd-session[7046]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:02.236622 systemd[1]: sshd@15-10.200.20.38:22-10.200.16.10:48750.service: Deactivated successfully. Mar 2 12:52:02.240107 systemd[1]: session-18.scope: Deactivated successfully. Mar 2 12:52:02.242023 systemd-logind[1870]: Session 18 logged out. Waiting for processes to exit. Mar 2 12:52:02.242957 systemd-logind[1870]: Removed session 18. Mar 2 12:52:02.322809 systemd[1]: Started sshd@16-10.200.20.38:22-10.200.16.10:34926.service - OpenSSH per-connection server daemon (10.200.16.10:34926). Mar 2 12:52:02.732655 sshd[7232]: Accepted publickey for core from 10.200.16.10 port 34926 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:52:02.733666 sshd-session[7232]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:52:02.737121 systemd-logind[1870]: New session 19 of user core. Mar 2 12:52:02.747606 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 2 12:52:03.077557 sshd[7235]: Connection closed by 10.200.16.10 port 34926 Mar 2 12:52:03.078249 sshd-session[7232]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:03.082233 systemd[1]: sshd@16-10.200.20.38:22-10.200.16.10:34926.service: Deactivated successfully. Mar 2 12:52:03.084701 systemd[1]: session-19.scope: Deactivated successfully. Mar 2 12:52:03.085689 systemd-logind[1870]: Session 19 logged out. Waiting for processes to exit. Mar 2 12:52:03.087826 systemd-logind[1870]: Removed session 19. Mar 2 12:52:03.167875 systemd[1]: Started sshd@17-10.200.20.38:22-10.200.16.10:34936.service - OpenSSH per-connection server daemon (10.200.16.10:34936). Mar 2 12:52:03.582293 sshd[7245]: Accepted publickey for core from 10.200.16.10 port 34936 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:52:03.583216 sshd-session[7245]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:52:03.587450 systemd-logind[1870]: New session 20 of user core. Mar 2 12:52:03.591611 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 2 12:52:03.857160 sshd[7272]: Connection closed by 10.200.16.10 port 34936 Mar 2 12:52:03.857086 sshd-session[7245]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:03.860980 systemd[1]: sshd@17-10.200.20.38:22-10.200.16.10:34936.service: Deactivated successfully. Mar 2 12:52:03.862434 systemd[1]: session-20.scope: Deactivated successfully. Mar 2 12:52:03.864360 systemd-logind[1870]: Session 20 logged out. Waiting for processes to exit. Mar 2 12:52:03.865374 systemd-logind[1870]: Removed session 20. Mar 2 12:52:08.956791 systemd[1]: Started sshd@18-10.200.20.38:22-10.200.16.10:34940.service - OpenSSH per-connection server daemon (10.200.16.10:34940). Mar 2 12:52:09.379014 sshd[7309]: Accepted publickey for core from 10.200.16.10 port 34940 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:52:09.379726 sshd-session[7309]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:52:09.385654 systemd-logind[1870]: New session 21 of user core. Mar 2 12:52:09.389711 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 2 12:52:09.651140 sshd[7312]: Connection closed by 10.200.16.10 port 34940 Mar 2 12:52:09.651681 sshd-session[7309]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:09.654525 systemd[1]: sshd@18-10.200.20.38:22-10.200.16.10:34940.service: Deactivated successfully. Mar 2 12:52:09.655943 systemd[1]: session-21.scope: Deactivated successfully. Mar 2 12:52:09.656656 systemd-logind[1870]: Session 21 logged out. Waiting for processes to exit. Mar 2 12:52:09.658330 systemd-logind[1870]: Removed session 21. Mar 2 12:52:14.738119 systemd[1]: Started sshd@19-10.200.20.38:22-10.200.16.10:47720.service - OpenSSH per-connection server daemon (10.200.16.10:47720). Mar 2 12:52:15.157544 sshd[7330]: Accepted publickey for core from 10.200.16.10 port 47720 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:52:15.158773 sshd-session[7330]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:52:15.162244 systemd-logind[1870]: New session 22 of user core. Mar 2 12:52:15.171599 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 2 12:52:15.437047 sshd[7333]: Connection closed by 10.200.16.10 port 47720 Mar 2 12:52:15.437573 sshd-session[7330]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:15.440605 systemd[1]: sshd@19-10.200.20.38:22-10.200.16.10:47720.service: Deactivated successfully. Mar 2 12:52:15.442334 systemd[1]: session-22.scope: Deactivated successfully. Mar 2 12:52:15.444538 systemd-logind[1870]: Session 22 logged out. Waiting for processes to exit. Mar 2 12:52:15.445485 systemd-logind[1870]: Removed session 22. Mar 2 12:52:20.526636 systemd[1]: Started sshd@20-10.200.20.38:22-10.200.16.10:40028.service - OpenSSH per-connection server daemon (10.200.16.10:40028). Mar 2 12:52:20.944612 sshd[7346]: Accepted publickey for core from 10.200.16.10 port 40028 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:52:20.945623 sshd-session[7346]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:52:20.949330 systemd-logind[1870]: New session 23 of user core. Mar 2 12:52:20.954602 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 2 12:52:21.219827 sshd[7349]: Connection closed by 10.200.16.10 port 40028 Mar 2 12:52:21.220185 sshd-session[7346]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:21.223879 systemd-logind[1870]: Session 23 logged out. Waiting for processes to exit. Mar 2 12:52:21.224549 systemd[1]: sshd@20-10.200.20.38:22-10.200.16.10:40028.service: Deactivated successfully. Mar 2 12:52:21.227159 systemd[1]: session-23.scope: Deactivated successfully. Mar 2 12:52:21.229034 systemd-logind[1870]: Removed session 23. Mar 2 12:52:26.313669 systemd[1]: Started sshd@21-10.200.20.38:22-10.200.16.10:40038.service - OpenSSH per-connection server daemon (10.200.16.10:40038). Mar 2 12:52:26.728638 sshd[7382]: Accepted publickey for core from 10.200.16.10 port 40038 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:52:26.729563 sshd-session[7382]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:52:26.733727 systemd-logind[1870]: New session 24 of user core. Mar 2 12:52:26.742603 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 2 12:52:27.015694 sshd[7385]: Connection closed by 10.200.16.10 port 40038 Mar 2 12:52:27.016281 sshd-session[7382]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:27.019577 systemd[1]: sshd@21-10.200.20.38:22-10.200.16.10:40038.service: Deactivated successfully. Mar 2 12:52:27.021629 systemd[1]: session-24.scope: Deactivated successfully. Mar 2 12:52:27.023067 systemd-logind[1870]: Session 24 logged out. Waiting for processes to exit. Mar 2 12:52:27.024928 systemd-logind[1870]: Removed session 24. Mar 2 12:52:32.108899 systemd[1]: Started sshd@22-10.200.20.38:22-10.200.16.10:48030.service - OpenSSH per-connection server daemon (10.200.16.10:48030). Mar 2 12:52:32.523039 sshd[7421]: Accepted publickey for core from 10.200.16.10 port 48030 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:52:32.523973 sshd-session[7421]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:52:32.528406 systemd-logind[1870]: New session 25 of user core. Mar 2 12:52:32.537617 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 2 12:52:32.796646 sshd[7424]: Connection closed by 10.200.16.10 port 48030 Mar 2 12:52:32.797195 sshd-session[7421]: pam_unix(sshd:session): session closed for user core Mar 2 12:52:32.799993 systemd[1]: sshd@22-10.200.20.38:22-10.200.16.10:48030.service: Deactivated successfully. Mar 2 12:52:32.802900 systemd[1]: session-25.scope: Deactivated successfully. Mar 2 12:52:32.804055 systemd-logind[1870]: Session 25 logged out. Waiting for processes to exit. Mar 2 12:52:32.804828 systemd-logind[1870]: Removed session 25.