Apr 28 00:14:42.881277 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Apr 28 00:14:42.881301 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Mon Apr 27 22:49:05 -00 2026 Apr 28 00:14:42.881312 kernel: KASLR enabled Apr 28 00:14:42.881318 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Apr 28 00:14:42.881324 kernel: efi: SMBIOS 3.0=0x139ed0000 MEMATTR=0x1390c1018 ACPI 2.0=0x136760018 RNG=0x13676e918 MEMRESERVE=0x136b43d18 Apr 28 00:14:42.881329 kernel: random: crng init done Apr 28 00:14:42.881336 kernel: ACPI: Early table checksum verification disabled Apr 28 00:14:42.881342 kernel: ACPI: RSDP 0x0000000136760018 000024 (v02 BOCHS ) Apr 28 00:14:42.881349 kernel: ACPI: XSDT 0x000000013676FE98 00006C (v01 BOCHS BXPC 00000001 01000013) Apr 28 00:14:42.881356 kernel: ACPI: FACP 0x000000013676FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881362 kernel: ACPI: DSDT 0x0000000136767518 001468 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881368 kernel: ACPI: APIC 0x000000013676FC18 000108 (v04 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881374 kernel: ACPI: PPTT 0x000000013676FD98 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881381 kernel: ACPI: GTDT 0x000000013676D898 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881388 kernel: ACPI: MCFG 0x000000013676FF98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881396 kernel: ACPI: SPCR 0x000000013676E818 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881402 kernel: ACPI: DBG2 0x000000013676E898 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881408 kernel: ACPI: IORT 0x000000013676E418 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 28 00:14:42.881415 kernel: ACPI: BGRT 0x000000013676E798 000038 (v01 INTEL EDK2 00000002 01000013) Apr 28 00:14:42.881421 kernel: ACPI: SPCR: console: pl011,mmio32,0x9000000,9600 Apr 28 00:14:42.881427 kernel: NUMA: Failed to initialise from firmware Apr 28 00:14:42.881434 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x0000000139ffffff] Apr 28 00:14:42.881440 kernel: NUMA: NODE_DATA [mem 0x13966f800-0x139674fff] Apr 28 00:14:42.881446 kernel: Zone ranges: Apr 28 00:14:42.881453 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Apr 28 00:14:42.881461 kernel: DMA32 empty Apr 28 00:14:42.881467 kernel: Normal [mem 0x0000000100000000-0x0000000139ffffff] Apr 28 00:14:42.881474 kernel: Movable zone start for each node Apr 28 00:14:42.881480 kernel: Early memory node ranges Apr 28 00:14:42.881486 kernel: node 0: [mem 0x0000000040000000-0x000000013676ffff] Apr 28 00:14:42.881493 kernel: node 0: [mem 0x0000000136770000-0x0000000136b3ffff] Apr 28 00:14:42.881499 kernel: node 0: [mem 0x0000000136b40000-0x0000000139e1ffff] Apr 28 00:14:42.881506 kernel: node 0: [mem 0x0000000139e20000-0x0000000139eaffff] Apr 28 00:14:42.881512 kernel: node 0: [mem 0x0000000139eb0000-0x0000000139ebffff] Apr 28 00:14:42.881518 kernel: node 0: [mem 0x0000000139ec0000-0x0000000139fdffff] Apr 28 00:14:42.881525 kernel: node 0: [mem 0x0000000139fe0000-0x0000000139ffffff] Apr 28 00:14:42.881531 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x0000000139ffffff] Apr 28 00:14:42.881539 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Apr 28 00:14:42.881546 kernel: psci: probing for conduit method from ACPI. Apr 28 00:14:42.881553 kernel: psci: PSCIv1.1 detected in firmware. Apr 28 00:14:42.881562 kernel: psci: Using standard PSCI v0.2 function IDs Apr 28 00:14:42.881569 kernel: psci: Trusted OS migration not required Apr 28 00:14:42.881576 kernel: psci: SMC Calling Convention v1.1 Apr 28 00:14:42.881584 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Apr 28 00:14:42.881591 kernel: percpu: Embedded 30 pages/cpu s85736 r8192 d28952 u122880 Apr 28 00:14:42.881598 kernel: pcpu-alloc: s85736 r8192 d28952 u122880 alloc=30*4096 Apr 28 00:14:42.881605 kernel: pcpu-alloc: [0] 0 [0] 1 Apr 28 00:14:42.881611 kernel: Detected PIPT I-cache on CPU0 Apr 28 00:14:42.881618 kernel: CPU features: detected: GIC system register CPU interface Apr 28 00:14:42.881625 kernel: CPU features: detected: Hardware dirty bit management Apr 28 00:14:42.881632 kernel: CPU features: detected: Spectre-v4 Apr 28 00:14:42.881638 kernel: CPU features: detected: Spectre-BHB Apr 28 00:14:42.881645 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 28 00:14:42.881653 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 28 00:14:42.881660 kernel: CPU features: detected: ARM erratum 1418040 Apr 28 00:14:42.881667 kernel: CPU features: detected: SSBS not fully self-synchronizing Apr 28 00:14:42.881674 kernel: alternatives: applying boot alternatives Apr 28 00:14:42.881682 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=5fbd74e24c605bcd6049a4229047ecffba5884416be782935a76f3959939199f Apr 28 00:14:42.881689 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 28 00:14:42.881697 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 28 00:14:42.881703 kernel: Fallback order for Node 0: 0 Apr 28 00:14:42.881827 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1008000 Apr 28 00:14:42.881836 kernel: Policy zone: Normal Apr 28 00:14:42.881843 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 28 00:14:42.881855 kernel: software IO TLB: area num 2. Apr 28 00:14:42.881862 kernel: software IO TLB: mapped [mem 0x00000000fbfff000-0x00000000fffff000] (64MB) Apr 28 00:14:42.881870 kernel: Memory: 3882816K/4096000K available (10304K kernel code, 2180K rwdata, 8116K rodata, 39424K init, 897K bss, 213184K reserved, 0K cma-reserved) Apr 28 00:14:42.881886 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 28 00:14:42.881894 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 28 00:14:42.881902 kernel: rcu: RCU event tracing is enabled. Apr 28 00:14:42.881909 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 28 00:14:42.881916 kernel: Trampoline variant of Tasks RCU enabled. Apr 28 00:14:42.881923 kernel: Tracing variant of Tasks RCU enabled. Apr 28 00:14:42.881930 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 28 00:14:42.881937 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 28 00:14:42.881943 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 28 00:14:42.881953 kernel: GICv3: 256 SPIs implemented Apr 28 00:14:42.881959 kernel: GICv3: 0 Extended SPIs implemented Apr 28 00:14:42.881966 kernel: Root IRQ handler: gic_handle_irq Apr 28 00:14:42.881973 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Apr 28 00:14:42.881980 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Apr 28 00:14:42.881987 kernel: ITS [mem 0x08080000-0x0809ffff] Apr 28 00:14:42.881994 kernel: ITS@0x0000000008080000: allocated 8192 Devices @1000c0000 (indirect, esz 8, psz 64K, shr 1) Apr 28 00:14:42.882001 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @1000d0000 (flat, esz 8, psz 64K, shr 1) Apr 28 00:14:42.882008 kernel: GICv3: using LPI property table @0x00000001000e0000 Apr 28 00:14:42.882015 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000001000f0000 Apr 28 00:14:42.882022 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 28 00:14:42.882030 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 28 00:14:42.882037 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Apr 28 00:14:42.882044 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Apr 28 00:14:42.882051 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Apr 28 00:14:42.882058 kernel: Console: colour dummy device 80x25 Apr 28 00:14:42.882065 kernel: ACPI: Core revision 20230628 Apr 28 00:14:42.882073 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Apr 28 00:14:42.882080 kernel: pid_max: default: 32768 minimum: 301 Apr 28 00:14:42.882087 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 28 00:14:42.882094 kernel: landlock: Up and running. Apr 28 00:14:42.882103 kernel: SELinux: Initializing. Apr 28 00:14:42.882110 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 28 00:14:42.882118 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 28 00:14:42.882125 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 28 00:14:42.882132 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 28 00:14:42.882139 kernel: rcu: Hierarchical SRCU implementation. Apr 28 00:14:42.882146 kernel: rcu: Max phase no-delay instances is 400. Apr 28 00:14:42.882154 kernel: Platform MSI: ITS@0x8080000 domain created Apr 28 00:14:42.882161 kernel: PCI/MSI: ITS@0x8080000 domain created Apr 28 00:14:42.882169 kernel: Remapping and enabling EFI services. Apr 28 00:14:42.882176 kernel: smp: Bringing up secondary CPUs ... Apr 28 00:14:42.882183 kernel: Detected PIPT I-cache on CPU1 Apr 28 00:14:42.882195 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Apr 28 00:14:42.882203 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000100100000 Apr 28 00:14:42.882210 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 28 00:14:42.882217 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Apr 28 00:14:42.882224 kernel: smp: Brought up 1 node, 2 CPUs Apr 28 00:14:42.882231 kernel: SMP: Total of 2 processors activated. Apr 28 00:14:42.882240 kernel: CPU features: detected: 32-bit EL0 Support Apr 28 00:14:42.882247 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 28 00:14:42.882255 kernel: CPU features: detected: Common not Private translations Apr 28 00:14:42.882268 kernel: CPU features: detected: CRC32 instructions Apr 28 00:14:42.882277 kernel: CPU features: detected: Enhanced Virtualization Traps Apr 28 00:14:42.882284 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 28 00:14:42.882291 kernel: CPU features: detected: LSE atomic instructions Apr 28 00:14:42.882299 kernel: CPU features: detected: Privileged Access Never Apr 28 00:14:42.882307 kernel: CPU features: detected: RAS Extension Support Apr 28 00:14:42.882316 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Apr 28 00:14:42.882324 kernel: CPU: All CPU(s) started at EL1 Apr 28 00:14:42.882331 kernel: alternatives: applying system-wide alternatives Apr 28 00:14:42.882338 kernel: devtmpfs: initialized Apr 28 00:14:42.882346 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 28 00:14:42.882354 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 28 00:14:42.882361 kernel: pinctrl core: initialized pinctrl subsystem Apr 28 00:14:42.882369 kernel: SMBIOS 3.0.0 present. Apr 28 00:14:42.882377 kernel: DMI: Hetzner vServer/KVM Virtual Machine, BIOS 20171111 11/11/2017 Apr 28 00:14:42.882385 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 28 00:14:42.882393 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Apr 28 00:14:42.882400 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 28 00:14:42.882408 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 28 00:14:42.882416 kernel: audit: initializing netlink subsys (disabled) Apr 28 00:14:42.882423 kernel: audit: type=2000 audit(0.014:1): state=initialized audit_enabled=0 res=1 Apr 28 00:14:42.882431 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 28 00:14:42.882438 kernel: cpuidle: using governor menu Apr 28 00:14:42.882447 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 28 00:14:42.882455 kernel: ASID allocator initialised with 32768 entries Apr 28 00:14:42.882462 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 28 00:14:42.882470 kernel: Serial: AMBA PL011 UART driver Apr 28 00:14:42.882478 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Apr 28 00:14:42.882485 kernel: Modules: 0 pages in range for non-PLT usage Apr 28 00:14:42.882493 kernel: Modules: 509008 pages in range for PLT usage Apr 28 00:14:42.882501 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 28 00:14:42.882508 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Apr 28 00:14:42.882517 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Apr 28 00:14:42.882525 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Apr 28 00:14:42.882533 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 28 00:14:42.882540 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Apr 28 00:14:42.882548 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Apr 28 00:14:42.882555 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Apr 28 00:14:42.882563 kernel: ACPI: Added _OSI(Module Device) Apr 28 00:14:42.882570 kernel: ACPI: Added _OSI(Processor Device) Apr 28 00:14:42.882577 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 28 00:14:42.882587 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 28 00:14:42.882594 kernel: ACPI: Interpreter enabled Apr 28 00:14:42.882602 kernel: ACPI: Using GIC for interrupt routing Apr 28 00:14:42.882609 kernel: ACPI: MCFG table detected, 1 entries Apr 28 00:14:42.882617 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Apr 28 00:14:42.882624 kernel: printk: console [ttyAMA0] enabled Apr 28 00:14:42.882631 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 28 00:14:42.883459 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 28 00:14:42.883581 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Apr 28 00:14:42.883653 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Apr 28 00:14:42.883718 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Apr 28 00:14:42.883849 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Apr 28 00:14:42.883862 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Apr 28 00:14:42.883870 kernel: PCI host bridge to bus 0000:00 Apr 28 00:14:42.883965 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Apr 28 00:14:42.884032 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Apr 28 00:14:42.884092 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Apr 28 00:14:42.884151 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 28 00:14:42.884236 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Apr 28 00:14:42.884319 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x038000 Apr 28 00:14:42.884388 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x11289000-0x11289fff] Apr 28 00:14:42.884456 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000600000-0x8000603fff 64bit pref] Apr 28 00:14:42.884536 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.884603 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x11288000-0x11288fff] Apr 28 00:14:42.884677 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.884789 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x11287000-0x11287fff] Apr 28 00:14:42.884870 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.884987 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x11286000-0x11286fff] Apr 28 00:14:42.885069 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.885138 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x11285000-0x11285fff] Apr 28 00:14:42.885534 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.885614 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x11284000-0x11284fff] Apr 28 00:14:42.885700 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.885795 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x11283000-0x11283fff] Apr 28 00:14:42.885874 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.885962 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x11282000-0x11282fff] Apr 28 00:14:42.886055 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.886138 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x11281000-0x11281fff] Apr 28 00:14:42.886212 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Apr 28 00:14:42.886287 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x11280000-0x11280fff] Apr 28 00:14:42.886384 kernel: pci 0000:00:04.0: [1b36:0002] type 00 class 0x070002 Apr 28 00:14:42.886459 kernel: pci 0000:00:04.0: reg 0x10: [io 0x0000-0x0007] Apr 28 00:14:42.886537 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Apr 28 00:14:42.886608 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x11000000-0x11000fff] Apr 28 00:14:42.886677 kernel: pci 0000:01:00.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Apr 28 00:14:42.886761 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Apr 28 00:14:42.886845 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Apr 28 00:14:42.886945 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x10e00000-0x10e03fff 64bit] Apr 28 00:14:42.887030 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Apr 28 00:14:42.887102 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x10c00000-0x10c00fff] Apr 28 00:14:42.887172 kernel: pci 0000:03:00.0: reg 0x20: [mem 0x8000100000-0x8000103fff 64bit pref] Apr 28 00:14:42.887249 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Apr 28 00:14:42.887319 kernel: pci 0000:04:00.0: reg 0x20: [mem 0x8000200000-0x8000203fff 64bit pref] Apr 28 00:14:42.887402 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Apr 28 00:14:42.887473 kernel: pci 0000:05:00.0: reg 0x14: [mem 0x10800000-0x10800fff] Apr 28 00:14:42.887551 kernel: pci 0000:05:00.0: reg 0x20: [mem 0x8000300000-0x8000303fff 64bit pref] Apr 28 00:14:42.887635 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Apr 28 00:14:42.887706 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x10600000-0x10600fff] Apr 28 00:14:42.891321 kernel: pci 0000:06:00.0: reg 0x20: [mem 0x8000400000-0x8000403fff 64bit pref] Apr 28 00:14:42.891439 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Apr 28 00:14:42.891535 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x10400000-0x10400fff] Apr 28 00:14:42.891609 kernel: pci 0000:07:00.0: reg 0x20: [mem 0x8000500000-0x8000503fff 64bit pref] Apr 28 00:14:42.891680 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Apr 28 00:14:42.891771 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 28 00:14:42.891854 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 01] add_size 100000 add_align 100000 Apr 28 00:14:42.891954 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff] to [bus 01] add_size 100000 add_align 100000 Apr 28 00:14:42.892042 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 28 00:14:42.892149 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 28 00:14:42.892252 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x001fffff] to [bus 02] add_size 100000 add_align 100000 Apr 28 00:14:42.892340 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 28 00:14:42.892413 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 03] add_size 100000 add_align 100000 Apr 28 00:14:42.892498 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Apr 28 00:14:42.892588 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 28 00:14:42.892656 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 04] add_size 100000 add_align 100000 Apr 28 00:14:42.892728 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 28 00:14:42.893236 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Apr 28 00:14:42.893418 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 05] add_size 100000 add_align 100000 Apr 28 00:14:42.893493 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff] to [bus 05] add_size 100000 add_align 100000 Apr 28 00:14:42.893574 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Apr 28 00:14:42.893661 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 06] add_size 100000 add_align 100000 Apr 28 00:14:42.893734 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff] to [bus 06] add_size 100000 add_align 100000 Apr 28 00:14:42.894978 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Apr 28 00:14:42.895052 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 07] add_size 100000 add_align 100000 Apr 28 00:14:42.895119 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff] to [bus 07] add_size 100000 add_align 100000 Apr 28 00:14:42.895191 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Apr 28 00:14:42.895258 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 08] add_size 200000 add_align 100000 Apr 28 00:14:42.895324 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff] to [bus 08] add_size 200000 add_align 100000 Apr 28 00:14:42.895399 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Apr 28 00:14:42.895470 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 09] add_size 200000 add_align 100000 Apr 28 00:14:42.895546 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 09] add_size 200000 add_align 100000 Apr 28 00:14:42.895620 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Apr 28 00:14:42.896814 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x8000000000-0x80001fffff 64bit pref] Apr 28 00:14:42.896960 kernel: pci 0000:00:02.1: BAR 14: assigned [mem 0x10200000-0x103fffff] Apr 28 00:14:42.897041 kernel: pci 0000:00:02.1: BAR 15: assigned [mem 0x8000200000-0x80003fffff 64bit pref] Apr 28 00:14:42.897115 kernel: pci 0000:00:02.2: BAR 14: assigned [mem 0x10400000-0x105fffff] Apr 28 00:14:42.897193 kernel: pci 0000:00:02.2: BAR 15: assigned [mem 0x8000400000-0x80005fffff 64bit pref] Apr 28 00:14:42.897315 kernel: pci 0000:00:02.3: BAR 14: assigned [mem 0x10600000-0x107fffff] Apr 28 00:14:42.897387 kernel: pci 0000:00:02.3: BAR 15: assigned [mem 0x8000600000-0x80007fffff 64bit pref] Apr 28 00:14:42.897456 kernel: pci 0000:00:02.4: BAR 14: assigned [mem 0x10800000-0x109fffff] Apr 28 00:14:42.897523 kernel: pci 0000:00:02.4: BAR 15: assigned [mem 0x8000800000-0x80009fffff 64bit pref] Apr 28 00:14:42.897595 kernel: pci 0000:00:02.5: BAR 14: assigned [mem 0x10a00000-0x10bfffff] Apr 28 00:14:42.897661 kernel: pci 0000:00:02.5: BAR 15: assigned [mem 0x8000a00000-0x8000bfffff 64bit pref] Apr 28 00:14:42.897735 kernel: pci 0000:00:02.6: BAR 14: assigned [mem 0x10c00000-0x10dfffff] Apr 28 00:14:42.897835 kernel: pci 0000:00:02.6: BAR 15: assigned [mem 0x8000c00000-0x8000dfffff 64bit pref] Apr 28 00:14:42.897922 kernel: pci 0000:00:02.7: BAR 14: assigned [mem 0x10e00000-0x10ffffff] Apr 28 00:14:42.897995 kernel: pci 0000:00:02.7: BAR 15: assigned [mem 0x8000e00000-0x8000ffffff 64bit pref] Apr 28 00:14:42.898064 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x11000000-0x111fffff] Apr 28 00:14:42.898132 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x8001000000-0x80011fffff 64bit pref] Apr 28 00:14:42.898204 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8001200000-0x8001203fff 64bit pref] Apr 28 00:14:42.898276 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x11200000-0x11200fff] Apr 28 00:14:42.898345 kernel: pci 0000:00:02.0: BAR 0: assigned [mem 0x11201000-0x11201fff] Apr 28 00:14:42.898412 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Apr 28 00:14:42.898478 kernel: pci 0000:00:02.1: BAR 0: assigned [mem 0x11202000-0x11202fff] Apr 28 00:14:42.898545 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Apr 28 00:14:42.898611 kernel: pci 0000:00:02.2: BAR 0: assigned [mem 0x11203000-0x11203fff] Apr 28 00:14:42.898678 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Apr 28 00:14:42.900546 kernel: pci 0000:00:02.3: BAR 0: assigned [mem 0x11204000-0x11204fff] Apr 28 00:14:42.900659 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Apr 28 00:14:42.900731 kernel: pci 0000:00:02.4: BAR 0: assigned [mem 0x11205000-0x11205fff] Apr 28 00:14:42.900887 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Apr 28 00:14:42.900973 kernel: pci 0000:00:02.5: BAR 0: assigned [mem 0x11206000-0x11206fff] Apr 28 00:14:42.901041 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Apr 28 00:14:42.901109 kernel: pci 0000:00:02.6: BAR 0: assigned [mem 0x11207000-0x11207fff] Apr 28 00:14:42.901175 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Apr 28 00:14:42.901242 kernel: pci 0000:00:02.7: BAR 0: assigned [mem 0x11208000-0x11208fff] Apr 28 00:14:42.901315 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Apr 28 00:14:42.901384 kernel: pci 0000:00:03.0: BAR 0: assigned [mem 0x11209000-0x11209fff] Apr 28 00:14:42.901449 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x9000-0x9fff] Apr 28 00:14:42.901520 kernel: pci 0000:00:04.0: BAR 0: assigned [io 0xa000-0xa007] Apr 28 00:14:42.901596 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x10000000-0x1007ffff pref] Apr 28 00:14:42.901665 kernel: pci 0000:01:00.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Apr 28 00:14:42.901733 kernel: pci 0000:01:00.0: BAR 1: assigned [mem 0x10080000-0x10080fff] Apr 28 00:14:42.902649 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Apr 28 00:14:42.902733 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Apr 28 00:14:42.902974 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff] Apr 28 00:14:42.903046 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref] Apr 28 00:14:42.903121 kernel: pci 0000:02:00.0: BAR 0: assigned [mem 0x10200000-0x10203fff 64bit] Apr 28 00:14:42.903197 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Apr 28 00:14:42.903263 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Apr 28 00:14:42.903329 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff] Apr 28 00:14:42.903395 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref] Apr 28 00:14:42.903472 kernel: pci 0000:03:00.0: BAR 4: assigned [mem 0x8000400000-0x8000403fff 64bit pref] Apr 28 00:14:42.903552 kernel: pci 0000:03:00.0: BAR 1: assigned [mem 0x10400000-0x10400fff] Apr 28 00:14:42.903623 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Apr 28 00:14:42.903701 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Apr 28 00:14:42.903797 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff] Apr 28 00:14:42.903867 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref] Apr 28 00:14:42.903987 kernel: pci 0000:04:00.0: BAR 4: assigned [mem 0x8000600000-0x8000603fff 64bit pref] Apr 28 00:14:42.904060 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Apr 28 00:14:42.904126 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Apr 28 00:14:42.904192 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff] Apr 28 00:14:42.904260 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref] Apr 28 00:14:42.904334 kernel: pci 0000:05:00.0: BAR 4: assigned [mem 0x8000800000-0x8000803fff 64bit pref] Apr 28 00:14:42.904413 kernel: pci 0000:05:00.0: BAR 1: assigned [mem 0x10800000-0x10800fff] Apr 28 00:14:42.904490 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Apr 28 00:14:42.904559 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Apr 28 00:14:42.904626 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Apr 28 00:14:42.904693 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref] Apr 28 00:14:42.904797 kernel: pci 0000:06:00.0: BAR 4: assigned [mem 0x8000a00000-0x8000a03fff 64bit pref] Apr 28 00:14:42.904873 kernel: pci 0000:06:00.0: BAR 1: assigned [mem 0x10a00000-0x10a00fff] Apr 28 00:14:42.904964 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Apr 28 00:14:42.905039 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Apr 28 00:14:42.905105 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff] Apr 28 00:14:42.905171 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref] Apr 28 00:14:42.905247 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x10c00000-0x10c7ffff pref] Apr 28 00:14:42.905316 kernel: pci 0000:07:00.0: BAR 4: assigned [mem 0x8000c00000-0x8000c03fff 64bit pref] Apr 28 00:14:42.905385 kernel: pci 0000:07:00.0: BAR 1: assigned [mem 0x10c80000-0x10c80fff] Apr 28 00:14:42.905454 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Apr 28 00:14:42.905521 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Apr 28 00:14:42.905591 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff] Apr 28 00:14:42.905656 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref] Apr 28 00:14:42.905724 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Apr 28 00:14:42.905931 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Apr 28 00:14:42.906006 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff] Apr 28 00:14:42.906072 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref] Apr 28 00:14:42.906141 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Apr 28 00:14:42.906209 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff] Apr 28 00:14:42.906282 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff] Apr 28 00:14:42.906348 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref] Apr 28 00:14:42.906417 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Apr 28 00:14:42.906476 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Apr 28 00:14:42.906535 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Apr 28 00:14:42.906608 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Apr 28 00:14:42.906671 kernel: pci_bus 0000:01: resource 1 [mem 0x10000000-0x101fffff] Apr 28 00:14:42.906734 kernel: pci_bus 0000:01: resource 2 [mem 0x8000000000-0x80001fffff 64bit pref] Apr 28 00:14:42.906842 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x2fff] Apr 28 00:14:42.906951 kernel: pci_bus 0000:02: resource 1 [mem 0x10200000-0x103fffff] Apr 28 00:14:42.907018 kernel: pci_bus 0000:02: resource 2 [mem 0x8000200000-0x80003fffff 64bit pref] Apr 28 00:14:42.907104 kernel: pci_bus 0000:03: resource 0 [io 0x3000-0x3fff] Apr 28 00:14:42.907171 kernel: pci_bus 0000:03: resource 1 [mem 0x10400000-0x105fffff] Apr 28 00:14:42.907297 kernel: pci_bus 0000:03: resource 2 [mem 0x8000400000-0x80005fffff 64bit pref] Apr 28 00:14:42.907405 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Apr 28 00:14:42.907489 kernel: pci_bus 0000:04: resource 1 [mem 0x10600000-0x107fffff] Apr 28 00:14:42.907575 kernel: pci_bus 0000:04: resource 2 [mem 0x8000600000-0x80007fffff 64bit pref] Apr 28 00:14:42.907647 kernel: pci_bus 0000:05: resource 0 [io 0x5000-0x5fff] Apr 28 00:14:42.907715 kernel: pci_bus 0000:05: resource 1 [mem 0x10800000-0x109fffff] Apr 28 00:14:42.907868 kernel: pci_bus 0000:05: resource 2 [mem 0x8000800000-0x80009fffff 64bit pref] Apr 28 00:14:42.907967 kernel: pci_bus 0000:06: resource 0 [io 0x6000-0x6fff] Apr 28 00:14:42.908030 kernel: pci_bus 0000:06: resource 1 [mem 0x10a00000-0x10bfffff] Apr 28 00:14:42.908095 kernel: pci_bus 0000:06: resource 2 [mem 0x8000a00000-0x8000bfffff 64bit pref] Apr 28 00:14:42.908162 kernel: pci_bus 0000:07: resource 0 [io 0x7000-0x7fff] Apr 28 00:14:42.908277 kernel: pci_bus 0000:07: resource 1 [mem 0x10c00000-0x10dfffff] Apr 28 00:14:42.908377 kernel: pci_bus 0000:07: resource 2 [mem 0x8000c00000-0x8000dfffff 64bit pref] Apr 28 00:14:42.908460 kernel: pci_bus 0000:08: resource 0 [io 0x8000-0x8fff] Apr 28 00:14:42.908521 kernel: pci_bus 0000:08: resource 1 [mem 0x10e00000-0x10ffffff] Apr 28 00:14:42.908620 kernel: pci_bus 0000:08: resource 2 [mem 0x8000e00000-0x8000ffffff 64bit pref] Apr 28 00:14:42.908698 kernel: pci_bus 0000:09: resource 0 [io 0x9000-0x9fff] Apr 28 00:14:42.908839 kernel: pci_bus 0000:09: resource 1 [mem 0x11000000-0x111fffff] Apr 28 00:14:42.908930 kernel: pci_bus 0000:09: resource 2 [mem 0x8001000000-0x80011fffff 64bit pref] Apr 28 00:14:42.908943 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Apr 28 00:14:42.908951 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Apr 28 00:14:42.908960 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Apr 28 00:14:42.908968 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Apr 28 00:14:42.908975 kernel: iommu: Default domain type: Translated Apr 28 00:14:42.908984 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 28 00:14:42.908992 kernel: efivars: Registered efivars operations Apr 28 00:14:42.909002 kernel: vgaarb: loaded Apr 28 00:14:42.909010 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 28 00:14:42.909018 kernel: VFS: Disk quotas dquot_6.6.0 Apr 28 00:14:42.909026 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 28 00:14:42.909034 kernel: pnp: PnP ACPI init Apr 28 00:14:42.909113 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Apr 28 00:14:42.909125 kernel: pnp: PnP ACPI: found 1 devices Apr 28 00:14:42.909133 kernel: NET: Registered PF_INET protocol family Apr 28 00:14:42.909141 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 28 00:14:42.909152 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 28 00:14:42.909159 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 28 00:14:42.909168 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 28 00:14:42.909176 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 28 00:14:42.909184 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 28 00:14:42.909191 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 28 00:14:42.909199 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 28 00:14:42.909207 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 28 00:14:42.909283 kernel: pci 0000:02:00.0: enabling device (0000 -> 0002) Apr 28 00:14:42.909297 kernel: PCI: CLS 0 bytes, default 64 Apr 28 00:14:42.909305 kernel: kvm [1]: HYP mode not available Apr 28 00:14:42.909313 kernel: Initialise system trusted keyrings Apr 28 00:14:42.909321 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 28 00:14:42.909328 kernel: Key type asymmetric registered Apr 28 00:14:42.909336 kernel: Asymmetric key parser 'x509' registered Apr 28 00:14:42.909344 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Apr 28 00:14:42.909352 kernel: io scheduler mq-deadline registered Apr 28 00:14:42.909360 kernel: io scheduler kyber registered Apr 28 00:14:42.909370 kernel: io scheduler bfq registered Apr 28 00:14:42.909379 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Apr 28 00:14:42.909448 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 50 Apr 28 00:14:42.909516 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 50 Apr 28 00:14:42.909585 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.909657 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 51 Apr 28 00:14:42.909728 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 51 Apr 28 00:14:42.909822 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.909943 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 52 Apr 28 00:14:42.910030 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 52 Apr 28 00:14:42.910098 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.910190 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 53 Apr 28 00:14:42.910267 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 53 Apr 28 00:14:42.910334 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.910404 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 54 Apr 28 00:14:42.910471 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 54 Apr 28 00:14:42.910538 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.910607 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 55 Apr 28 00:14:42.910678 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 55 Apr 28 00:14:42.910757 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.910830 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 56 Apr 28 00:14:42.910914 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 56 Apr 28 00:14:42.910986 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.911057 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 57 Apr 28 00:14:42.911129 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 57 Apr 28 00:14:42.911199 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.911210 kernel: ACPI: \_SB_.PCI0.GSI3: Enabled at IRQ 38 Apr 28 00:14:42.911277 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 58 Apr 28 00:14:42.911346 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 58 Apr 28 00:14:42.911412 kernel: pcieport 0000:00:03.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 28 00:14:42.911423 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Apr 28 00:14:42.911434 kernel: ACPI: button: Power Button [PWRB] Apr 28 00:14:42.911442 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Apr 28 00:14:42.911515 kernel: virtio-pci 0000:04:00.0: enabling device (0000 -> 0002) Apr 28 00:14:42.911590 kernel: virtio-pci 0000:07:00.0: enabling device (0000 -> 0002) Apr 28 00:14:42.911601 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 28 00:14:42.911609 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Apr 28 00:14:42.911678 kernel: serial 0000:00:04.0: enabling device (0000 -> 0001) Apr 28 00:14:42.911689 kernel: 0000:00:04.0: ttyS0 at I/O 0xa000 (irq = 45, base_baud = 115200) is a 16550A Apr 28 00:14:42.911697 kernel: thunder_xcv, ver 1.0 Apr 28 00:14:42.911707 kernel: thunder_bgx, ver 1.0 Apr 28 00:14:42.911715 kernel: nicpf, ver 1.0 Apr 28 00:14:42.911723 kernel: nicvf, ver 1.0 Apr 28 00:14:42.911823 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 28 00:14:42.911948 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-04-28T00:14:42 UTC (1777335282) Apr 28 00:14:42.911961 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 28 00:14:42.911970 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Apr 28 00:14:42.911978 kernel: watchdog: Delayed init of the lockup detector failed: -19 Apr 28 00:14:42.911990 kernel: watchdog: Hard watchdog permanently disabled Apr 28 00:14:42.911998 kernel: NET: Registered PF_INET6 protocol family Apr 28 00:14:42.912006 kernel: Segment Routing with IPv6 Apr 28 00:14:42.912014 kernel: In-situ OAM (IOAM) with IPv6 Apr 28 00:14:42.912022 kernel: NET: Registered PF_PACKET protocol family Apr 28 00:14:42.912029 kernel: Key type dns_resolver registered Apr 28 00:14:42.912037 kernel: registered taskstats version 1 Apr 28 00:14:42.912045 kernel: Loading compiled-in X.509 certificates Apr 28 00:14:42.912053 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: 6c96a5ff031ece119b3ff0073294cdad6eea39a2' Apr 28 00:14:42.912063 kernel: Key type .fscrypt registered Apr 28 00:14:42.912071 kernel: Key type fscrypt-provisioning registered Apr 28 00:14:42.912078 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 28 00:14:42.912086 kernel: ima: Allocated hash algorithm: sha1 Apr 28 00:14:42.912094 kernel: ima: No architecture policies found Apr 28 00:14:42.912102 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 28 00:14:42.912109 kernel: clk: Disabling unused clocks Apr 28 00:14:42.912117 kernel: Freeing unused kernel memory: 39424K Apr 28 00:14:42.912125 kernel: Run /init as init process Apr 28 00:14:42.912134 kernel: with arguments: Apr 28 00:14:42.912142 kernel: /init Apr 28 00:14:42.912150 kernel: with environment: Apr 28 00:14:42.912157 kernel: HOME=/ Apr 28 00:14:42.912165 kernel: TERM=linux Apr 28 00:14:42.912175 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 28 00:14:42.912185 systemd[1]: Detected virtualization kvm. Apr 28 00:14:42.912193 systemd[1]: Detected architecture arm64. Apr 28 00:14:42.912203 systemd[1]: Running in initrd. Apr 28 00:14:42.912211 systemd[1]: No hostname configured, using default hostname. Apr 28 00:14:42.912219 systemd[1]: Hostname set to . Apr 28 00:14:42.912228 systemd[1]: Initializing machine ID from VM UUID. Apr 28 00:14:42.912236 systemd[1]: Queued start job for default target initrd.target. Apr 28 00:14:42.912244 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 28 00:14:42.912253 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 28 00:14:42.912262 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 28 00:14:42.912271 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 28 00:14:42.912280 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 28 00:14:42.912290 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 28 00:14:42.912300 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 28 00:14:42.912309 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 28 00:14:42.912317 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 28 00:14:42.912339 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 28 00:14:42.912352 systemd[1]: Reached target paths.target - Path Units. Apr 28 00:14:42.912361 systemd[1]: Reached target slices.target - Slice Units. Apr 28 00:14:42.912369 systemd[1]: Reached target swap.target - Swaps. Apr 28 00:14:42.912377 systemd[1]: Reached target timers.target - Timer Units. Apr 28 00:14:42.912385 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 28 00:14:42.912394 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 28 00:14:42.912402 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 28 00:14:42.912411 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 28 00:14:42.912421 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 28 00:14:42.912429 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 28 00:14:42.912438 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 28 00:14:42.912446 systemd[1]: Reached target sockets.target - Socket Units. Apr 28 00:14:42.912454 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 28 00:14:42.912463 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 28 00:14:42.912471 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 28 00:14:42.912479 systemd[1]: Starting systemd-fsck-usr.service... Apr 28 00:14:42.912487 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 28 00:14:42.912497 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 28 00:14:42.912506 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 28 00:14:42.912539 systemd-journald[237]: Collecting audit messages is disabled. Apr 28 00:14:42.912560 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 28 00:14:42.912571 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 28 00:14:42.912579 systemd[1]: Finished systemd-fsck-usr.service. Apr 28 00:14:42.912588 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 28 00:14:42.912596 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 28 00:14:42.912607 systemd-journald[237]: Journal started Apr 28 00:14:42.912627 systemd-journald[237]: Runtime Journal (/run/log/journal/1e43d220cdfa4a059a2a44cf67566f95) is 8.0M, max 76.6M, 68.6M free. Apr 28 00:14:42.892393 systemd-modules-load[238]: Inserted module 'overlay' Apr 28 00:14:42.913817 systemd[1]: Started systemd-journald.service - Journal Service. Apr 28 00:14:42.916862 kernel: Bridge firewalling registered Apr 28 00:14:42.917106 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 28 00:14:42.918307 systemd-modules-load[238]: Inserted module 'br_netfilter' Apr 28 00:14:42.919540 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 28 00:14:42.929967 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 28 00:14:42.935035 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 28 00:14:42.945969 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 28 00:14:42.947006 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 28 00:14:42.953011 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 28 00:14:42.962638 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 28 00:14:42.963760 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 28 00:14:42.968837 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 28 00:14:42.986127 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 28 00:14:42.991547 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 28 00:14:42.993258 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 28 00:14:43.004552 dracut-cmdline[271]: dracut-dracut-053 Apr 28 00:14:43.012462 dracut-cmdline[271]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=5fbd74e24c605bcd6049a4229047ecffba5884416be782935a76f3959939199f Apr 28 00:14:43.036872 systemd-resolved[273]: Positive Trust Anchors: Apr 28 00:14:43.037648 systemd-resolved[273]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 28 00:14:43.037682 systemd-resolved[273]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 28 00:14:43.047651 systemd-resolved[273]: Defaulting to hostname 'linux'. Apr 28 00:14:43.049707 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 28 00:14:43.051309 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 28 00:14:43.082823 kernel: SCSI subsystem initialized Apr 28 00:14:43.086784 kernel: Loading iSCSI transport class v2.0-870. Apr 28 00:14:43.095143 kernel: iscsi: registered transport (tcp) Apr 28 00:14:43.108831 kernel: iscsi: registered transport (qla4xxx) Apr 28 00:14:43.108896 kernel: QLogic iSCSI HBA Driver Apr 28 00:14:43.157500 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 28 00:14:43.165081 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 28 00:14:43.185128 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 28 00:14:43.185193 kernel: device-mapper: uevent: version 1.0.3 Apr 28 00:14:43.185205 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 28 00:14:43.236803 kernel: raid6: neonx8 gen() 15647 MB/s Apr 28 00:14:43.253810 kernel: raid6: neonx4 gen() 15487 MB/s Apr 28 00:14:43.270823 kernel: raid6: neonx2 gen() 13187 MB/s Apr 28 00:14:43.287834 kernel: raid6: neonx1 gen() 10429 MB/s Apr 28 00:14:43.304795 kernel: raid6: int64x8 gen() 6905 MB/s Apr 28 00:14:43.321819 kernel: raid6: int64x4 gen() 7305 MB/s Apr 28 00:14:43.338816 kernel: raid6: int64x2 gen() 6095 MB/s Apr 28 00:14:43.355823 kernel: raid6: int64x1 gen() 5037 MB/s Apr 28 00:14:43.355934 kernel: raid6: using algorithm neonx8 gen() 15647 MB/s Apr 28 00:14:43.372809 kernel: raid6: .... xor() 11922 MB/s, rmw enabled Apr 28 00:14:43.372901 kernel: raid6: using neon recovery algorithm Apr 28 00:14:43.377780 kernel: xor: measuring software checksum speed Apr 28 00:14:43.377846 kernel: 8regs : 19797 MB/sec Apr 28 00:14:43.377869 kernel: 32regs : 18215 MB/sec Apr 28 00:14:43.378858 kernel: arm64_neon : 26981 MB/sec Apr 28 00:14:43.378908 kernel: xor: using function: arm64_neon (26981 MB/sec) Apr 28 00:14:43.429799 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 28 00:14:43.443198 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 28 00:14:43.448982 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 28 00:14:43.464170 systemd-udevd[455]: Using default interface naming scheme 'v255'. Apr 28 00:14:43.468484 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 28 00:14:43.477950 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 28 00:14:43.495076 dracut-pre-trigger[464]: rd.md=0: removing MD RAID activation Apr 28 00:14:43.534606 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 28 00:14:43.540105 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 28 00:14:43.589678 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 28 00:14:43.597120 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 28 00:14:43.618774 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 28 00:14:43.622227 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 28 00:14:43.624557 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 28 00:14:43.626461 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 28 00:14:43.631955 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 28 00:14:43.655896 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 28 00:14:43.683006 kernel: scsi host0: Virtio SCSI HBA Apr 28 00:14:43.691611 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU CD-ROM 2.5+ PQ: 0 ANSI: 5 Apr 28 00:14:43.691689 kernel: scsi 0:0:0:1: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Apr 28 00:14:43.722570 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 28 00:14:43.722697 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 28 00:14:43.725521 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 28 00:14:43.727630 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 28 00:14:43.728086 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 28 00:14:43.731540 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 28 00:14:43.739958 kernel: ACPI: bus type USB registered Apr 28 00:14:43.740020 kernel: usbcore: registered new interface driver usbfs Apr 28 00:14:43.740761 kernel: usbcore: registered new interface driver hub Apr 28 00:14:43.740781 kernel: usbcore: registered new device driver usb Apr 28 00:14:43.745297 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 28 00:14:43.755774 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 28 00:14:43.756035 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Apr 28 00:14:43.756136 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Apr 28 00:14:43.757850 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 28 00:14:43.758026 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Apr 28 00:14:43.758112 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Apr 28 00:14:43.758193 kernel: hub 1-0:1.0: USB hub found Apr 28 00:14:43.758294 kernel: hub 1-0:1.0: 4 ports detected Apr 28 00:14:43.761774 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Apr 28 00:14:43.765729 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 28 00:14:43.771768 kernel: hub 2-0:1.0: USB hub found Apr 28 00:14:43.772967 kernel: hub 2-0:1.0: 4 ports detected Apr 28 00:14:43.775013 kernel: sr 0:0:0:0: Power-on or device reset occurred Apr 28 00:14:43.777066 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 28 00:14:43.782250 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 16x/50x cd/rw xa/form2 cdda tray Apr 28 00:14:43.782439 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 28 00:14:43.783765 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Apr 28 00:14:43.795823 kernel: sd 0:0:0:1: Power-on or device reset occurred Apr 28 00:14:43.796048 kernel: sd 0:0:0:1: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Apr 28 00:14:43.796138 kernel: sd 0:0:0:1: [sda] Write Protect is off Apr 28 00:14:43.796228 kernel: sd 0:0:0:1: [sda] Mode Sense: 63 00 00 08 Apr 28 00:14:43.797038 kernel: sd 0:0:0:1: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Apr 28 00:14:43.803947 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 28 00:14:43.804021 kernel: GPT:17805311 != 80003071 Apr 28 00:14:43.804034 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 28 00:14:43.804045 kernel: GPT:17805311 != 80003071 Apr 28 00:14:43.805174 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 28 00:14:43.805209 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 28 00:14:43.806768 kernel: sd 0:0:0:1: [sda] Attached SCSI disk Apr 28 00:14:43.809463 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 28 00:14:43.848767 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/sda6 scanned by (udev-worker) (513) Apr 28 00:14:43.860797 kernel: BTRFS: device fsid 4ceb9780-605b-47f7-8c1f-b3fcb9f87ddc devid 1 transid 32 /dev/sda3 scanned by (udev-worker) (501) Apr 28 00:14:43.861654 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Apr 28 00:14:43.875239 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Apr 28 00:14:43.882196 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 28 00:14:43.887476 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Apr 28 00:14:43.888359 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Apr 28 00:14:43.897042 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 28 00:14:43.905608 disk-uuid[571]: Primary Header is updated. Apr 28 00:14:43.905608 disk-uuid[571]: Secondary Entries is updated. Apr 28 00:14:43.905608 disk-uuid[571]: Secondary Header is updated. Apr 28 00:14:43.914780 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 28 00:14:43.921813 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 28 00:14:44.003844 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Apr 28 00:14:44.138602 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input1 Apr 28 00:14:44.138685 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Apr 28 00:14:44.139028 kernel: usbcore: registered new interface driver usbhid Apr 28 00:14:44.139789 kernel: usbhid: USB HID core driver Apr 28 00:14:44.247084 kernel: usb 1-2: new high-speed USB device number 3 using xhci_hcd Apr 28 00:14:44.377785 kernel: input: QEMU QEMU USB Keyboard as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-2/1-2:1.0/0003:0627:0001.0002/input/input2 Apr 28 00:14:44.431796 kernel: hid-generic 0003:0627:0001.0002: input,hidraw1: USB HID v1.11 Keyboard [QEMU QEMU USB Keyboard] on usb-0000:02:00.0-2/input0 Apr 28 00:14:44.928964 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 28 00:14:44.929018 disk-uuid[572]: The operation has completed successfully. Apr 28 00:14:44.979213 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 28 00:14:44.979338 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 28 00:14:44.996047 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 28 00:14:45.002789 sh[590]: Success Apr 28 00:14:45.019999 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Apr 28 00:14:45.083442 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 28 00:14:45.087199 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 28 00:14:45.090100 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 28 00:14:45.114343 kernel: BTRFS info (device dm-0): first mount of filesystem 4ceb9780-605b-47f7-8c1f-b3fcb9f87ddc Apr 28 00:14:45.114423 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Apr 28 00:14:45.114452 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 28 00:14:45.114490 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 28 00:14:45.115450 kernel: BTRFS info (device dm-0): using free space tree Apr 28 00:14:45.121794 kernel: BTRFS info (device dm-0): enabling ssd optimizations Apr 28 00:14:45.124071 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 28 00:14:45.126357 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 28 00:14:45.133046 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 28 00:14:45.138098 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 28 00:14:45.149197 kernel: BTRFS info (device sda6): first mount of filesystem 57367c84-0f72-4cbc-90cb-9cf0a8258220 Apr 28 00:14:45.149252 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 28 00:14:45.149265 kernel: BTRFS info (device sda6): using free space tree Apr 28 00:14:45.161192 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 28 00:14:45.161262 kernel: BTRFS info (device sda6): auto enabling async discard Apr 28 00:14:45.172333 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 28 00:14:45.174804 kernel: BTRFS info (device sda6): last unmount of filesystem 57367c84-0f72-4cbc-90cb-9cf0a8258220 Apr 28 00:14:45.180916 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 28 00:14:45.186023 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 28 00:14:45.263274 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 28 00:14:45.274913 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 28 00:14:45.289501 ignition[688]: Ignition 2.19.0 Apr 28 00:14:45.289511 ignition[688]: Stage: fetch-offline Apr 28 00:14:45.289551 ignition[688]: no configs at "/usr/lib/ignition/base.d" Apr 28 00:14:45.289559 ignition[688]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 28 00:14:45.293226 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 28 00:14:45.289715 ignition[688]: parsed url from cmdline: "" Apr 28 00:14:45.289718 ignition[688]: no config URL provided Apr 28 00:14:45.289722 ignition[688]: reading system config file "/usr/lib/ignition/user.ign" Apr 28 00:14:45.289730 ignition[688]: no config at "/usr/lib/ignition/user.ign" Apr 28 00:14:45.289734 ignition[688]: failed to fetch config: resource requires networking Apr 28 00:14:45.300448 systemd-networkd[776]: lo: Link UP Apr 28 00:14:45.290076 ignition[688]: Ignition finished successfully Apr 28 00:14:45.300453 systemd-networkd[776]: lo: Gained carrier Apr 28 00:14:45.302700 systemd-networkd[776]: Enumeration completed Apr 28 00:14:45.303135 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 28 00:14:45.303593 systemd-networkd[776]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:45.303597 systemd-networkd[776]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 28 00:14:45.304677 systemd[1]: Reached target network.target - Network. Apr 28 00:14:45.304998 systemd-networkd[776]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:45.305001 systemd-networkd[776]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 28 00:14:45.306378 systemd-networkd[776]: eth0: Link UP Apr 28 00:14:45.306382 systemd-networkd[776]: eth0: Gained carrier Apr 28 00:14:45.306391 systemd-networkd[776]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:45.310151 systemd-networkd[776]: eth1: Link UP Apr 28 00:14:45.310157 systemd-networkd[776]: eth1: Gained carrier Apr 28 00:14:45.310170 systemd-networkd[776]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:45.312319 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 28 00:14:45.329872 ignition[780]: Ignition 2.19.0 Apr 28 00:14:45.329884 ignition[780]: Stage: fetch Apr 28 00:14:45.330141 ignition[780]: no configs at "/usr/lib/ignition/base.d" Apr 28 00:14:45.330155 ignition[780]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 28 00:14:45.330273 ignition[780]: parsed url from cmdline: "" Apr 28 00:14:45.330277 ignition[780]: no config URL provided Apr 28 00:14:45.330283 ignition[780]: reading system config file "/usr/lib/ignition/user.ign" Apr 28 00:14:45.330292 ignition[780]: no config at "/usr/lib/ignition/user.ign" Apr 28 00:14:45.330314 ignition[780]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Apr 28 00:14:45.331634 ignition[780]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Apr 28 00:14:45.350886 systemd-networkd[776]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 28 00:14:45.368884 systemd-networkd[776]: eth0: DHCPv4 address 178.105.21.241/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 28 00:14:45.531719 ignition[780]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Apr 28 00:14:45.537087 ignition[780]: GET result: OK Apr 28 00:14:45.537226 ignition[780]: parsing config with SHA512: 4d86128e601e5ade36c22851a93596a383cd6f44bf8a363a5e35c71f1f066fef072f3132a67696f4cd041886bc2f3474cc8a952ef56b7fc2f6c0e70d7e382cc2 Apr 28 00:14:45.544772 unknown[780]: fetched base config from "system" Apr 28 00:14:45.544786 unknown[780]: fetched base config from "system" Apr 28 00:14:45.545510 ignition[780]: fetch: fetch complete Apr 28 00:14:45.544794 unknown[780]: fetched user config from "hetzner" Apr 28 00:14:45.545515 ignition[780]: fetch: fetch passed Apr 28 00:14:45.547694 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 28 00:14:45.545568 ignition[780]: Ignition finished successfully Apr 28 00:14:45.555046 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 28 00:14:45.567831 ignition[787]: Ignition 2.19.0 Apr 28 00:14:45.567842 ignition[787]: Stage: kargs Apr 28 00:14:45.568038 ignition[787]: no configs at "/usr/lib/ignition/base.d" Apr 28 00:14:45.568048 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 28 00:14:45.569023 ignition[787]: kargs: kargs passed Apr 28 00:14:45.571638 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 28 00:14:45.569098 ignition[787]: Ignition finished successfully Apr 28 00:14:45.580003 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 28 00:14:45.596778 ignition[793]: Ignition 2.19.0 Apr 28 00:14:45.596789 ignition[793]: Stage: disks Apr 28 00:14:45.596996 ignition[793]: no configs at "/usr/lib/ignition/base.d" Apr 28 00:14:45.597007 ignition[793]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 28 00:14:45.600366 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 28 00:14:45.598039 ignition[793]: disks: disks passed Apr 28 00:14:45.602811 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 28 00:14:45.598095 ignition[793]: Ignition finished successfully Apr 28 00:14:45.603816 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 28 00:14:45.605217 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 28 00:14:45.606271 systemd[1]: Reached target sysinit.target - System Initialization. Apr 28 00:14:45.607406 systemd[1]: Reached target basic.target - Basic System. Apr 28 00:14:45.613069 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 28 00:14:45.632344 systemd-fsck[801]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Apr 28 00:14:45.639656 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 28 00:14:45.646969 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 28 00:14:45.696794 kernel: EXT4-fs (sda9): mounted filesystem 2d8f83b6-5f3b-4fc5-b0f6-3405e8e67f7b r/w with ordered data mode. Quota mode: none. Apr 28 00:14:45.696440 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 28 00:14:45.698089 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 28 00:14:45.705902 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 28 00:14:45.711307 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 28 00:14:45.713782 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 28 00:14:45.714572 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 28 00:14:45.714605 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 28 00:14:45.726061 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 28 00:14:45.728672 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by mount (809) Apr 28 00:14:45.734025 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 28 00:14:45.738578 kernel: BTRFS info (device sda6): first mount of filesystem 57367c84-0f72-4cbc-90cb-9cf0a8258220 Apr 28 00:14:45.738637 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 28 00:14:45.738657 kernel: BTRFS info (device sda6): using free space tree Apr 28 00:14:45.749027 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 28 00:14:45.749097 kernel: BTRFS info (device sda6): auto enabling async discard Apr 28 00:14:45.759301 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 28 00:14:45.787340 initrd-setup-root[836]: cut: /sysroot/etc/passwd: No such file or directory Apr 28 00:14:45.793437 coreos-metadata[811]: Apr 28 00:14:45.793 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Apr 28 00:14:45.796965 coreos-metadata[811]: Apr 28 00:14:45.794 INFO Fetch successful Apr 28 00:14:45.798261 coreos-metadata[811]: Apr 28 00:14:45.798 INFO wrote hostname ci-4081-3-7-n-51c70c830d to /sysroot/etc/hostname Apr 28 00:14:45.799452 initrd-setup-root[843]: cut: /sysroot/etc/group: No such file or directory Apr 28 00:14:45.800445 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 28 00:14:45.806963 initrd-setup-root[851]: cut: /sysroot/etc/shadow: No such file or directory Apr 28 00:14:45.815028 initrd-setup-root[858]: cut: /sysroot/etc/gshadow: No such file or directory Apr 28 00:14:45.914244 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 28 00:14:45.924623 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 28 00:14:45.929832 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 28 00:14:45.935780 kernel: BTRFS info (device sda6): last unmount of filesystem 57367c84-0f72-4cbc-90cb-9cf0a8258220 Apr 28 00:14:45.961077 ignition[926]: INFO : Ignition 2.19.0 Apr 28 00:14:45.961077 ignition[926]: INFO : Stage: mount Apr 28 00:14:45.961077 ignition[926]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 28 00:14:45.961077 ignition[926]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 28 00:14:45.963069 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 28 00:14:45.967137 ignition[926]: INFO : mount: mount passed Apr 28 00:14:45.967137 ignition[926]: INFO : Ignition finished successfully Apr 28 00:14:45.964185 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 28 00:14:45.969961 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 28 00:14:46.115255 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 28 00:14:46.123113 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 28 00:14:46.134205 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (937) Apr 28 00:14:46.134277 kernel: BTRFS info (device sda6): first mount of filesystem 57367c84-0f72-4cbc-90cb-9cf0a8258220 Apr 28 00:14:46.135971 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 28 00:14:46.136020 kernel: BTRFS info (device sda6): using free space tree Apr 28 00:14:46.139766 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 28 00:14:46.139862 kernel: BTRFS info (device sda6): auto enabling async discard Apr 28 00:14:46.143767 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 28 00:14:46.168035 ignition[954]: INFO : Ignition 2.19.0 Apr 28 00:14:46.168035 ignition[954]: INFO : Stage: files Apr 28 00:14:46.168035 ignition[954]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 28 00:14:46.168035 ignition[954]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 28 00:14:46.171862 ignition[954]: DEBUG : files: compiled without relabeling support, skipping Apr 28 00:14:46.171862 ignition[954]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 28 00:14:46.171862 ignition[954]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 28 00:14:46.176283 ignition[954]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 28 00:14:46.176283 ignition[954]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 28 00:14:46.178350 ignition[954]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 28 00:14:46.176937 unknown[954]: wrote ssh authorized keys file for user: core Apr 28 00:14:46.180372 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Apr 28 00:14:46.180372 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Apr 28 00:14:46.261863 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 28 00:14:46.335779 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Apr 28 00:14:46.335779 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 28 00:14:46.335779 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 28 00:14:46.335779 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 28 00:14:46.342283 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-arm64.raw: attempt #1 Apr 28 00:14:46.664873 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 28 00:14:47.191047 systemd-networkd[776]: eth1: Gained IPv6LL Apr 28 00:14:47.255277 systemd-networkd[776]: eth0: Gained IPv6LL Apr 28 00:14:47.316207 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 28 00:14:47.316207 ignition[954]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Apr 28 00:14:47.320757 ignition[954]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 28 00:14:47.320757 ignition[954]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 28 00:14:47.320757 ignition[954]: INFO : files: files passed Apr 28 00:14:47.320757 ignition[954]: INFO : Ignition finished successfully Apr 28 00:14:47.325446 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 28 00:14:47.339400 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 28 00:14:47.343200 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 28 00:14:47.349302 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 28 00:14:47.350987 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 28 00:14:47.362908 initrd-setup-root-after-ignition[982]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 28 00:14:47.362908 initrd-setup-root-after-ignition[982]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 28 00:14:47.366513 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 28 00:14:47.370858 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 28 00:14:47.372081 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 28 00:14:47.383088 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 28 00:14:47.412960 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 28 00:14:47.413094 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 28 00:14:47.417544 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 28 00:14:47.418769 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 28 00:14:47.420478 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 28 00:14:47.428063 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 28 00:14:47.451132 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 28 00:14:47.462104 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 28 00:14:47.474125 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 28 00:14:47.475704 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 28 00:14:47.476622 systemd[1]: Stopped target timers.target - Timer Units. Apr 28 00:14:47.477876 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 28 00:14:47.478090 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 28 00:14:47.479695 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 28 00:14:47.481089 systemd[1]: Stopped target basic.target - Basic System. Apr 28 00:14:47.483040 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 28 00:14:47.484758 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 28 00:14:47.486880 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 28 00:14:47.488838 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 28 00:14:47.490452 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 28 00:14:47.491816 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 28 00:14:47.493081 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 28 00:14:47.494122 systemd[1]: Stopped target swap.target - Swaps. Apr 28 00:14:47.495069 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 28 00:14:47.495235 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 28 00:14:47.496560 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 28 00:14:47.497714 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 28 00:14:47.498871 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 28 00:14:47.498983 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 28 00:14:47.500297 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 28 00:14:47.500462 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 28 00:14:47.502107 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 28 00:14:47.502272 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 28 00:14:47.503505 systemd[1]: ignition-files.service: Deactivated successfully. Apr 28 00:14:47.503652 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 28 00:14:47.504580 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 28 00:14:47.504712 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 28 00:14:47.510085 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 28 00:14:47.510680 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 28 00:14:47.512001 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 28 00:14:47.519915 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 28 00:14:47.521123 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 28 00:14:47.521410 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 28 00:14:47.526191 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 28 00:14:47.526957 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 28 00:14:47.539517 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 28 00:14:47.540732 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 28 00:14:47.541503 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 28 00:14:47.543705 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 28 00:14:47.545945 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 28 00:14:47.550030 ignition[1006]: INFO : Ignition 2.19.0 Apr 28 00:14:47.550030 ignition[1006]: INFO : Stage: umount Apr 28 00:14:47.551063 ignition[1006]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 28 00:14:47.551063 ignition[1006]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 28 00:14:47.553548 ignition[1006]: INFO : umount: umount passed Apr 28 00:14:47.553548 ignition[1006]: INFO : Ignition finished successfully Apr 28 00:14:47.552772 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 28 00:14:47.552903 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 28 00:14:47.554632 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 28 00:14:47.554712 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 28 00:14:47.556164 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 28 00:14:47.556208 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 28 00:14:47.557425 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 28 00:14:47.557464 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 28 00:14:47.558712 systemd[1]: Stopped target network.target - Network. Apr 28 00:14:47.559779 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 28 00:14:47.559843 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 28 00:14:47.561306 systemd[1]: Stopped target paths.target - Path Units. Apr 28 00:14:47.562286 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 28 00:14:47.566103 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 28 00:14:47.568294 systemd[1]: Stopped target slices.target - Slice Units. Apr 28 00:14:47.570086 systemd[1]: Stopped target sockets.target - Socket Units. Apr 28 00:14:47.571435 systemd[1]: iscsid.socket: Deactivated successfully. Apr 28 00:14:47.571484 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 28 00:14:47.572392 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 28 00:14:47.572428 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 28 00:14:47.573387 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 28 00:14:47.573445 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 28 00:14:47.574491 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 28 00:14:47.574536 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 28 00:14:47.575683 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 28 00:14:47.575727 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 28 00:14:47.576932 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 28 00:14:47.578139 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 28 00:14:47.580793 systemd-networkd[776]: eth0: DHCPv6 lease lost Apr 28 00:14:47.587876 systemd-networkd[776]: eth1: DHCPv6 lease lost Apr 28 00:14:47.588200 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 28 00:14:47.588346 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 28 00:14:47.591577 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 28 00:14:47.591731 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 28 00:14:47.595283 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 28 00:14:47.595343 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 28 00:14:47.604008 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 28 00:14:47.606890 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 28 00:14:47.606962 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 28 00:14:47.608391 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 28 00:14:47.608443 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 28 00:14:47.610021 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 28 00:14:47.610069 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 28 00:14:47.611634 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 28 00:14:47.611683 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 28 00:14:47.614330 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 28 00:14:47.626716 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 28 00:14:47.626977 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 28 00:14:47.637024 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 28 00:14:47.637255 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 28 00:14:47.642442 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 28 00:14:47.642512 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 28 00:14:47.643549 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 28 00:14:47.643582 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 28 00:14:47.644909 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 28 00:14:47.645011 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 28 00:14:47.647287 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 28 00:14:47.647368 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 28 00:14:47.649552 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 28 00:14:47.649621 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 28 00:14:47.660990 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 28 00:14:47.663439 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 28 00:14:47.663552 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 28 00:14:47.666190 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 28 00:14:47.666285 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 28 00:14:47.677284 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 28 00:14:47.677523 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 28 00:14:47.679721 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 28 00:14:47.692849 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 28 00:14:47.702706 systemd[1]: Switching root. Apr 28 00:14:47.738898 systemd-journald[237]: Journal stopped Apr 28 00:14:48.649581 systemd-journald[237]: Received SIGTERM from PID 1 (systemd). Apr 28 00:14:48.649641 kernel: SELinux: policy capability network_peer_controls=1 Apr 28 00:14:48.649654 kernel: SELinux: policy capability open_perms=1 Apr 28 00:14:48.649666 kernel: SELinux: policy capability extended_socket_class=1 Apr 28 00:14:48.649676 kernel: SELinux: policy capability always_check_network=0 Apr 28 00:14:48.649686 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 28 00:14:48.649695 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 28 00:14:48.649704 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 28 00:14:48.649715 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 28 00:14:48.649725 kernel: audit: type=1403 audit(1777335287.907:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 28 00:14:48.649735 systemd[1]: Successfully loaded SELinux policy in 37.568ms. Apr 28 00:14:48.651862 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 11.113ms. Apr 28 00:14:48.651889 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 28 00:14:48.651901 systemd[1]: Detected virtualization kvm. Apr 28 00:14:48.651912 systemd[1]: Detected architecture arm64. Apr 28 00:14:48.651922 systemd[1]: Detected first boot. Apr 28 00:14:48.651933 systemd[1]: Hostname set to . Apr 28 00:14:48.651943 systemd[1]: Initializing machine ID from VM UUID. Apr 28 00:14:48.651958 zram_generator::config[1048]: No configuration found. Apr 28 00:14:48.651973 systemd[1]: Populated /etc with preset unit settings. Apr 28 00:14:48.651984 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 28 00:14:48.651994 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 28 00:14:48.652004 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 28 00:14:48.652015 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 28 00:14:48.652026 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 28 00:14:48.652042 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 28 00:14:48.652052 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 28 00:14:48.652062 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 28 00:14:48.652074 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 28 00:14:48.652085 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 28 00:14:48.652095 systemd[1]: Created slice user.slice - User and Session Slice. Apr 28 00:14:48.652109 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 28 00:14:48.652121 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 28 00:14:48.652131 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 28 00:14:48.652142 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 28 00:14:48.652156 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 28 00:14:48.652172 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 28 00:14:48.652183 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Apr 28 00:14:48.652193 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 28 00:14:48.652203 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 28 00:14:48.652213 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 28 00:14:48.652223 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 28 00:14:48.652234 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 28 00:14:48.652246 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 28 00:14:48.652256 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 28 00:14:48.652267 systemd[1]: Reached target slices.target - Slice Units. Apr 28 00:14:48.652277 systemd[1]: Reached target swap.target - Swaps. Apr 28 00:14:48.652287 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 28 00:14:48.652297 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 28 00:14:48.652308 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 28 00:14:48.652318 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 28 00:14:48.652328 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 28 00:14:48.652341 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 28 00:14:48.652351 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 28 00:14:48.652362 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 28 00:14:48.652372 systemd[1]: Mounting media.mount - External Media Directory... Apr 28 00:14:48.652382 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 28 00:14:48.652392 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 28 00:14:48.652403 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 28 00:14:48.652414 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 28 00:14:48.652424 systemd[1]: Reached target machines.target - Containers. Apr 28 00:14:48.652436 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 28 00:14:48.652447 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 28 00:14:48.652457 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 28 00:14:48.652468 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 28 00:14:48.652478 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 28 00:14:48.652492 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 28 00:14:48.652505 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 28 00:14:48.652515 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 28 00:14:48.652525 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 28 00:14:48.652536 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 28 00:14:48.652547 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 28 00:14:48.652557 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 28 00:14:48.652568 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 28 00:14:48.652581 systemd[1]: Stopped systemd-fsck-usr.service. Apr 28 00:14:48.652591 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 28 00:14:48.652602 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 28 00:14:48.652612 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 28 00:14:48.652622 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 28 00:14:48.652633 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 28 00:14:48.652644 systemd[1]: verity-setup.service: Deactivated successfully. Apr 28 00:14:48.652654 systemd[1]: Stopped verity-setup.service. Apr 28 00:14:48.652669 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 28 00:14:48.652679 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 28 00:14:48.652691 systemd[1]: Mounted media.mount - External Media Directory. Apr 28 00:14:48.652701 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 28 00:14:48.652757 systemd-journald[1115]: Collecting audit messages is disabled. Apr 28 00:14:48.652783 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 28 00:14:48.652795 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 28 00:14:48.652808 systemd-journald[1115]: Journal started Apr 28 00:14:48.652843 systemd-journald[1115]: Runtime Journal (/run/log/journal/1e43d220cdfa4a059a2a44cf67566f95) is 8.0M, max 76.6M, 68.6M free. Apr 28 00:14:48.402478 systemd[1]: Queued start job for default target multi-user.target. Apr 28 00:14:48.426083 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Apr 28 00:14:48.426545 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 28 00:14:48.655004 systemd[1]: Started systemd-journald.service - Journal Service. Apr 28 00:14:48.655754 kernel: ACPI: bus type drm_connector registered Apr 28 00:14:48.662019 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 28 00:14:48.664473 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 28 00:14:48.664643 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 28 00:14:48.666173 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 28 00:14:48.669837 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 28 00:14:48.671957 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 28 00:14:48.675085 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 28 00:14:48.676103 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 28 00:14:48.676268 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 28 00:14:48.677755 kernel: fuse: init (API version 7.39) Apr 28 00:14:48.677787 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 28 00:14:48.682095 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 28 00:14:48.684612 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 28 00:14:48.686768 kernel: loop: module loaded Apr 28 00:14:48.684778 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 28 00:14:48.686002 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 28 00:14:48.689691 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 28 00:14:48.690671 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 28 00:14:48.707558 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 28 00:14:48.717863 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 28 00:14:48.721547 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 28 00:14:48.724303 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 28 00:14:48.724352 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 28 00:14:48.729632 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Apr 28 00:14:48.736995 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 28 00:14:48.745005 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 28 00:14:48.746125 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 28 00:14:48.749193 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 28 00:14:48.751072 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 28 00:14:48.751901 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 28 00:14:48.754983 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 28 00:14:48.757861 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 28 00:14:48.761024 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 28 00:14:48.762931 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 28 00:14:48.766782 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 28 00:14:48.767896 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 28 00:14:48.768737 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 28 00:14:48.771891 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 28 00:14:48.786952 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 28 00:14:48.800863 systemd-journald[1115]: Time spent on flushing to /var/log/journal/1e43d220cdfa4a059a2a44cf67566f95 is 89.780ms for 1123 entries. Apr 28 00:14:48.800863 systemd-journald[1115]: System Journal (/var/log/journal/1e43d220cdfa4a059a2a44cf67566f95) is 8.0M, max 584.8M, 576.8M free. Apr 28 00:14:48.914336 systemd-journald[1115]: Received client request to flush runtime journal. Apr 28 00:14:48.914380 kernel: loop0: detected capacity change from 0 to 209336 Apr 28 00:14:48.914394 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 28 00:14:48.805808 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 28 00:14:48.808431 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 28 00:14:48.819037 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 28 00:14:48.839161 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 28 00:14:48.851073 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 28 00:14:48.928797 kernel: loop1: detected capacity change from 0 to 114432 Apr 28 00:14:48.883209 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 28 00:14:48.884654 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 28 00:14:48.887976 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Apr 28 00:14:48.899032 udevadm[1173]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Apr 28 00:14:48.911920 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 28 00:14:48.920587 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 28 00:14:48.923397 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 28 00:14:48.961647 systemd-tmpfiles[1178]: ACLs are not supported, ignoring. Apr 28 00:14:48.962094 systemd-tmpfiles[1178]: ACLs are not supported, ignoring. Apr 28 00:14:48.965775 kernel: loop2: detected capacity change from 0 to 114328 Apr 28 00:14:48.971365 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 28 00:14:48.997531 kernel: loop3: detected capacity change from 0 to 8 Apr 28 00:14:49.016829 kernel: loop4: detected capacity change from 0 to 209336 Apr 28 00:14:49.035772 kernel: loop5: detected capacity change from 0 to 114432 Apr 28 00:14:49.050785 kernel: loop6: detected capacity change from 0 to 114328 Apr 28 00:14:49.073771 kernel: loop7: detected capacity change from 0 to 8 Apr 28 00:14:49.074398 (sd-merge)[1188]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Apr 28 00:14:49.075097 (sd-merge)[1188]: Merged extensions into '/usr'. Apr 28 00:14:49.086073 systemd[1]: Reloading requested from client PID 1162 ('systemd-sysext') (unit systemd-sysext.service)... Apr 28 00:14:49.086089 systemd[1]: Reloading... Apr 28 00:14:49.220800 zram_generator::config[1214]: No configuration found. Apr 28 00:14:49.235265 ldconfig[1157]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 28 00:14:49.346629 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 28 00:14:49.403394 systemd[1]: Reloading finished in 316 ms. Apr 28 00:14:49.426499 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 28 00:14:49.428094 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 28 00:14:49.439041 systemd[1]: Starting ensure-sysext.service... Apr 28 00:14:49.442119 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 28 00:14:49.457538 systemd[1]: Reloading requested from client PID 1252 ('systemctl') (unit ensure-sysext.service)... Apr 28 00:14:49.457559 systemd[1]: Reloading... Apr 28 00:14:49.484596 systemd-tmpfiles[1253]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 28 00:14:49.488615 systemd-tmpfiles[1253]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 28 00:14:49.489522 systemd-tmpfiles[1253]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 28 00:14:49.489842 systemd-tmpfiles[1253]: ACLs are not supported, ignoring. Apr 28 00:14:49.489900 systemd-tmpfiles[1253]: ACLs are not supported, ignoring. Apr 28 00:14:49.493336 systemd-tmpfiles[1253]: Detected autofs mount point /boot during canonicalization of boot. Apr 28 00:14:49.493351 systemd-tmpfiles[1253]: Skipping /boot Apr 28 00:14:49.509493 systemd-tmpfiles[1253]: Detected autofs mount point /boot during canonicalization of boot. Apr 28 00:14:49.509515 systemd-tmpfiles[1253]: Skipping /boot Apr 28 00:14:49.566769 zram_generator::config[1292]: No configuration found. Apr 28 00:14:49.651358 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 28 00:14:49.698113 systemd[1]: Reloading finished in 240 ms. Apr 28 00:14:49.719244 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 28 00:14:49.728709 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 28 00:14:49.741114 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 28 00:14:49.746188 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 28 00:14:49.753133 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 28 00:14:49.755794 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 28 00:14:49.762603 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 28 00:14:49.768847 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 28 00:14:49.771844 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 28 00:14:49.777118 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 28 00:14:49.782195 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 28 00:14:49.785585 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 28 00:14:49.787958 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 28 00:14:49.792024 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 28 00:14:49.796105 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 28 00:14:49.796255 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 28 00:14:49.799059 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 28 00:14:49.807174 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 28 00:14:49.808531 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 28 00:14:49.810852 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 28 00:14:49.817168 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 28 00:14:49.818785 systemd[1]: Finished ensure-sysext.service. Apr 28 00:14:49.832091 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 28 00:14:49.837109 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 28 00:14:49.841840 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 28 00:14:49.843631 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 28 00:14:49.845878 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 28 00:14:49.847341 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 28 00:14:49.848433 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 28 00:14:49.850568 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 28 00:14:49.850716 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 28 00:14:49.853993 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 28 00:14:49.854154 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 28 00:14:49.860112 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 28 00:14:49.860206 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 28 00:14:49.860244 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 28 00:14:49.863831 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 28 00:14:49.880487 systemd-udevd[1328]: Using default interface naming scheme 'v255'. Apr 28 00:14:49.891212 augenrules[1356]: No rules Apr 28 00:14:49.895852 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 28 00:14:49.906393 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 28 00:14:49.930975 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 28 00:14:49.938029 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 28 00:14:50.000915 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 28 00:14:50.001737 systemd[1]: Reached target time-set.target - System Time Set. Apr 28 00:14:50.067175 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Apr 28 00:14:50.069124 systemd-resolved[1326]: Positive Trust Anchors: Apr 28 00:14:50.069427 systemd-resolved[1326]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 28 00:14:50.069524 systemd-resolved[1326]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 28 00:14:50.079473 systemd-resolved[1326]: Using system hostname 'ci-4081-3-7-n-51c70c830d'. Apr 28 00:14:50.082231 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 28 00:14:50.082914 systemd-networkd[1371]: lo: Link UP Apr 28 00:14:50.083043 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 28 00:14:50.083234 systemd-networkd[1371]: lo: Gained carrier Apr 28 00:14:50.084042 systemd-networkd[1371]: Enumeration completed Apr 28 00:14:50.085645 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 28 00:14:50.086547 systemd[1]: Reached target network.target - Network. Apr 28 00:14:50.110571 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 28 00:14:50.166064 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (1383) Apr 28 00:14:50.211463 kernel: mousedev: PS/2 mouse device common for all mice Apr 28 00:14:50.211905 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Apr 28 00:14:50.213340 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 28 00:14:50.227460 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 28 00:14:50.231288 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 28 00:14:50.234568 systemd-networkd[1371]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:50.235337 systemd-networkd[1371]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 28 00:14:50.236997 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 28 00:14:50.237142 systemd-networkd[1371]: eth0: Link UP Apr 28 00:14:50.237145 systemd-networkd[1371]: eth0: Gained carrier Apr 28 00:14:50.237164 systemd-networkd[1371]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:50.237796 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 28 00:14:50.237843 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 28 00:14:50.238198 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 28 00:14:50.238763 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 28 00:14:50.252175 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 28 00:14:50.253227 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 28 00:14:50.254065 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 28 00:14:50.258736 systemd-networkd[1371]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:50.258818 systemd-networkd[1371]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 28 00:14:50.261091 systemd-networkd[1371]: eth1: Link UP Apr 28 00:14:50.261217 systemd-networkd[1371]: eth1: Gained carrier Apr 28 00:14:50.261239 systemd-networkd[1371]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 28 00:14:50.261820 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 28 00:14:50.262953 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 28 00:14:50.263890 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 28 00:14:50.297109 systemd-networkd[1371]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 28 00:14:50.297936 systemd-timesyncd[1341]: Network configuration changed, trying to establish connection. Apr 28 00:14:50.306085 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 28 00:14:50.306907 systemd-networkd[1371]: eth0: DHCPv4 address 178.105.21.241/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 28 00:14:50.308023 systemd-timesyncd[1341]: Network configuration changed, trying to establish connection. Apr 28 00:14:50.314530 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 28 00:14:50.316594 kernel: [drm] pci: virtio-gpu-pci detected at 0000:00:01.0 Apr 28 00:14:50.316664 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Apr 28 00:14:50.316677 kernel: [drm] features: -context_init Apr 28 00:14:50.319989 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 28 00:14:50.320764 kernel: [drm] number of scanouts: 1 Apr 28 00:14:50.320826 kernel: [drm] number of cap sets: 0 Apr 28 00:14:50.332893 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Apr 28 00:14:50.340854 kernel: Console: switching to colour frame buffer device 160x50 Apr 28 00:14:50.347780 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Apr 28 00:14:50.352409 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 28 00:14:50.355324 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 28 00:14:50.355497 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 28 00:14:50.363076 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 28 00:14:50.426227 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 28 00:14:50.459724 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 28 00:14:50.467030 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 28 00:14:50.490592 lvm[1436]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 28 00:14:50.518765 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 28 00:14:50.520789 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 28 00:14:50.521679 systemd[1]: Reached target sysinit.target - System Initialization. Apr 28 00:14:50.522651 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 28 00:14:50.523478 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 28 00:14:50.524562 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 28 00:14:50.525494 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 28 00:14:50.526342 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 28 00:14:50.527079 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 28 00:14:50.527117 systemd[1]: Reached target paths.target - Path Units. Apr 28 00:14:50.527598 systemd[1]: Reached target timers.target - Timer Units. Apr 28 00:14:50.530841 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 28 00:14:50.534322 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 28 00:14:50.539728 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 28 00:14:50.541953 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 28 00:14:50.543539 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 28 00:14:50.544506 systemd[1]: Reached target sockets.target - Socket Units. Apr 28 00:14:50.545205 systemd[1]: Reached target basic.target - Basic System. Apr 28 00:14:50.545861 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 28 00:14:50.545890 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 28 00:14:50.548911 systemd[1]: Starting containerd.service - containerd container runtime... Apr 28 00:14:50.552761 lvm[1441]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 28 00:14:50.553127 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 28 00:14:50.559033 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 28 00:14:50.568590 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 28 00:14:50.575305 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 28 00:14:50.576119 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 28 00:14:50.580868 jq[1445]: false Apr 28 00:14:50.577375 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 28 00:14:50.588118 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 28 00:14:50.589839 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Apr 28 00:14:50.592820 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 28 00:14:50.596016 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 28 00:14:50.607097 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 28 00:14:50.608549 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 28 00:14:50.610123 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 28 00:14:50.612322 systemd[1]: Starting update-engine.service - Update Engine... Apr 28 00:14:50.616898 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 28 00:14:50.620166 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 28 00:14:50.626126 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 28 00:14:50.626872 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 28 00:14:50.629591 jq[1456]: true Apr 28 00:14:50.639260 dbus-daemon[1444]: [system] SELinux support is enabled Apr 28 00:14:50.639442 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 28 00:14:50.645313 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 28 00:14:50.645363 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 28 00:14:50.646179 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 28 00:14:50.646197 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 28 00:14:50.654813 jq[1463]: true Apr 28 00:14:50.660519 extend-filesystems[1446]: Found loop4 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found loop5 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found loop6 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found loop7 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda1 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda2 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda3 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found usr Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda4 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda6 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda7 Apr 28 00:14:50.671951 extend-filesystems[1446]: Found sda9 Apr 28 00:14:50.671951 extend-filesystems[1446]: Checking size of /dev/sda9 Apr 28 00:14:50.725966 extend-filesystems[1446]: Resized partition /dev/sda9 Apr 28 00:14:50.679566 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 28 00:14:50.726702 coreos-metadata[1443]: Apr 28 00:14:50.704 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Apr 28 00:14:50.726702 coreos-metadata[1443]: Apr 28 00:14:50.723 INFO Fetch successful Apr 28 00:14:50.726702 coreos-metadata[1443]: Apr 28 00:14:50.723 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Apr 28 00:14:50.726702 coreos-metadata[1443]: Apr 28 00:14:50.724 INFO Fetch successful Apr 28 00:14:50.733550 extend-filesystems[1486]: resize2fs 1.47.1 (20-May-2024) Apr 28 00:14:50.743490 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Apr 28 00:14:50.679768 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 28 00:14:50.743623 update_engine[1455]: I20260428 00:14:50.716478 1455 main.cc:92] Flatcar Update Engine starting Apr 28 00:14:50.743623 update_engine[1455]: I20260428 00:14:50.725730 1455 update_check_scheduler.cc:74] Next update check in 3m40s Apr 28 00:14:50.707720 (ntainerd)[1481]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 28 00:14:50.744172 tar[1459]: linux-arm64/LICENSE Apr 28 00:14:50.744172 tar[1459]: linux-arm64/helm Apr 28 00:14:50.708789 systemd[1]: motdgen.service: Deactivated successfully. Apr 28 00:14:50.709095 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 28 00:14:50.726037 systemd[1]: Started update-engine.service - Update Engine. Apr 28 00:14:50.736373 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 28 00:14:50.780708 systemd-logind[1454]: New seat seat0. Apr 28 00:14:50.783108 systemd-logind[1454]: Watching system buttons on /dev/input/event0 (Power Button) Apr 28 00:14:50.783124 systemd-logind[1454]: Watching system buttons on /dev/input/event2 (QEMU QEMU USB Keyboard) Apr 28 00:14:50.783333 systemd[1]: Started systemd-logind.service - User Login Management. Apr 28 00:14:50.855910 bash[1506]: Updated "/home/core/.ssh/authorized_keys" Apr 28 00:14:50.857217 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 28 00:14:50.869129 systemd[1]: Starting sshkeys.service... Apr 28 00:14:50.901769 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (1382) Apr 28 00:14:50.909338 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Apr 28 00:14:50.934782 extend-filesystems[1486]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Apr 28 00:14:50.934782 extend-filesystems[1486]: old_desc_blocks = 1, new_desc_blocks = 5 Apr 28 00:14:50.934782 extend-filesystems[1486]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Apr 28 00:14:50.950823 extend-filesystems[1446]: Resized filesystem in /dev/sda9 Apr 28 00:14:50.950823 extend-filesystems[1446]: Found sr0 Apr 28 00:14:50.936515 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 28 00:14:50.937806 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 28 00:14:50.944177 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Apr 28 00:14:50.957266 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Apr 28 00:14:50.966924 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 28 00:14:50.979232 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 28 00:14:51.011503 containerd[1481]: time="2026-04-28T00:14:51.011085720Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Apr 28 00:14:51.067600 containerd[1481]: time="2026-04-28T00:14:51.067547360Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 28 00:14:51.069757 coreos-metadata[1522]: Apr 28 00:14:51.067 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069407760Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069445160Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069473320Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069649120Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069666720Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069725440Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069737880Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.069982320Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.070000920Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.070014640Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070047 containerd[1481]: time="2026-04-28T00:14:51.070025480Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070253 containerd[1481]: time="2026-04-28T00:14:51.070096920Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070934 containerd[1481]: time="2026-04-28T00:14:51.070284200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070934 containerd[1481]: time="2026-04-28T00:14:51.070478520Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 28 00:14:51.070934 containerd[1481]: time="2026-04-28T00:14:51.070497320Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 28 00:14:51.070934 containerd[1481]: time="2026-04-28T00:14:51.070632640Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 28 00:14:51.070934 containerd[1481]: time="2026-04-28T00:14:51.070678000Z" level=info msg="metadata content store policy set" policy=shared Apr 28 00:14:51.071885 coreos-metadata[1522]: Apr 28 00:14:51.071 INFO Fetch successful Apr 28 00:14:51.076046 unknown[1522]: wrote ssh authorized keys file for user: core Apr 28 00:14:51.080765 containerd[1481]: time="2026-04-28T00:14:51.079219000Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 28 00:14:51.080765 containerd[1481]: time="2026-04-28T00:14:51.079530000Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 28 00:14:51.080765 containerd[1481]: time="2026-04-28T00:14:51.079558520Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 28 00:14:51.080765 containerd[1481]: time="2026-04-28T00:14:51.079624680Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 28 00:14:51.080765 containerd[1481]: time="2026-04-28T00:14:51.079640200Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 28 00:14:51.080765 containerd[1481]: time="2026-04-28T00:14:51.079861080Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 28 00:14:51.082382 containerd[1481]: time="2026-04-28T00:14:51.082305200Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 28 00:14:51.082599 containerd[1481]: time="2026-04-28T00:14:51.082575000Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 28 00:14:51.082643 containerd[1481]: time="2026-04-28T00:14:51.082602960Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 28 00:14:51.082666 containerd[1481]: time="2026-04-28T00:14:51.082621120Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 28 00:14:51.082666 containerd[1481]: time="2026-04-28T00:14:51.082660880Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082727 containerd[1481]: time="2026-04-28T00:14:51.082682320Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082727 containerd[1481]: time="2026-04-28T00:14:51.082704560Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082727 containerd[1481]: time="2026-04-28T00:14:51.082723080Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082824 containerd[1481]: time="2026-04-28T00:14:51.082756800Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082824 containerd[1481]: time="2026-04-28T00:14:51.082773160Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082824 containerd[1481]: time="2026-04-28T00:14:51.082804080Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082882 containerd[1481]: time="2026-04-28T00:14:51.082823800Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 28 00:14:51.082882 containerd[1481]: time="2026-04-28T00:14:51.082851680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.082882 containerd[1481]: time="2026-04-28T00:14:51.082870240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.082934 containerd[1481]: time="2026-04-28T00:14:51.082886440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.082934 containerd[1481]: time="2026-04-28T00:14:51.082904840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.082934 containerd[1481]: time="2026-04-28T00:14:51.082920120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.082987 containerd[1481]: time="2026-04-28T00:14:51.082934600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.082987 containerd[1481]: time="2026-04-28T00:14:51.082951680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.082987 containerd[1481]: time="2026-04-28T00:14:51.082978400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083036 containerd[1481]: time="2026-04-28T00:14:51.082995840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083036 containerd[1481]: time="2026-04-28T00:14:51.083014800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083036 containerd[1481]: time="2026-04-28T00:14:51.083030280Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083087 containerd[1481]: time="2026-04-28T00:14:51.083045800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083087 containerd[1481]: time="2026-04-28T00:14:51.083062000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083120 containerd[1481]: time="2026-04-28T00:14:51.083082960Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 28 00:14:51.083120 containerd[1481]: time="2026-04-28T00:14:51.083113800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083228 containerd[1481]: time="2026-04-28T00:14:51.083129600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083228 containerd[1481]: time="2026-04-28T00:14:51.083168960Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 28 00:14:51.083329 containerd[1481]: time="2026-04-28T00:14:51.083308400Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 28 00:14:51.083356 containerd[1481]: time="2026-04-28T00:14:51.083344400Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 28 00:14:51.083377 containerd[1481]: time="2026-04-28T00:14:51.083361080Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 28 00:14:51.083395 containerd[1481]: time="2026-04-28T00:14:51.083377360Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 28 00:14:51.083395 containerd[1481]: time="2026-04-28T00:14:51.083388160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.083449 containerd[1481]: time="2026-04-28T00:14:51.083405560Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 28 00:14:51.083449 containerd[1481]: time="2026-04-28T00:14:51.083419560Z" level=info msg="NRI interface is disabled by configuration." Apr 28 00:14:51.083449 containerd[1481]: time="2026-04-28T00:14:51.083430000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 28 00:14:51.089826 containerd[1481]: time="2026-04-28T00:14:51.088120920Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 28 00:14:51.089826 containerd[1481]: time="2026-04-28T00:14:51.088220280Z" level=info msg="Connect containerd service" Apr 28 00:14:51.089826 containerd[1481]: time="2026-04-28T00:14:51.088293920Z" level=info msg="using legacy CRI server" Apr 28 00:14:51.089826 containerd[1481]: time="2026-04-28T00:14:51.088309920Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 28 00:14:51.089826 containerd[1481]: time="2026-04-28T00:14:51.088423760Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 28 00:14:51.092758 containerd[1481]: time="2026-04-28T00:14:51.090151560Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 28 00:14:51.092758 containerd[1481]: time="2026-04-28T00:14:51.092100960Z" level=info msg="Start subscribing containerd event" Apr 28 00:14:51.092758 containerd[1481]: time="2026-04-28T00:14:51.092153040Z" level=info msg="Start recovering state" Apr 28 00:14:51.091533 locksmithd[1488]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 28 00:14:51.097898 containerd[1481]: time="2026-04-28T00:14:51.092774640Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 28 00:14:51.097898 containerd[1481]: time="2026-04-28T00:14:51.093404400Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 28 00:14:51.097898 containerd[1481]: time="2026-04-28T00:14:51.093642920Z" level=info msg="Start event monitor" Apr 28 00:14:51.097898 containerd[1481]: time="2026-04-28T00:14:51.093657680Z" level=info msg="Start snapshots syncer" Apr 28 00:14:51.097898 containerd[1481]: time="2026-04-28T00:14:51.093668320Z" level=info msg="Start cni network conf syncer for default" Apr 28 00:14:51.097898 containerd[1481]: time="2026-04-28T00:14:51.093736800Z" level=info msg="Start streaming server" Apr 28 00:14:51.097898 containerd[1481]: time="2026-04-28T00:14:51.094656240Z" level=info msg="containerd successfully booted in 0.084423s" Apr 28 00:14:51.095253 systemd[1]: Started containerd.service - containerd container runtime. Apr 28 00:14:51.122952 update-ssh-keys[1534]: Updated "/home/core/.ssh/authorized_keys" Apr 28 00:14:51.126095 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Apr 28 00:14:51.133290 systemd[1]: Finished sshkeys.service. Apr 28 00:14:51.331507 sshd_keygen[1489]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 28 00:14:51.354092 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 28 00:14:51.365092 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 28 00:14:51.373096 systemd[1]: issuegen.service: Deactivated successfully. Apr 28 00:14:51.373307 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 28 00:14:51.383914 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 28 00:14:51.394502 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 28 00:14:51.403073 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 28 00:14:51.410897 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Apr 28 00:14:51.412503 systemd[1]: Reached target getty.target - Login Prompts. Apr 28 00:14:51.446346 tar[1459]: linux-arm64/README.md Apr 28 00:14:51.457218 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 28 00:14:51.542992 systemd-networkd[1371]: eth0: Gained IPv6LL Apr 28 00:14:51.544341 systemd-timesyncd[1341]: Network configuration changed, trying to establish connection. Apr 28 00:14:51.549592 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 28 00:14:51.552569 systemd[1]: Reached target network-online.target - Network is Online. Apr 28 00:14:51.563176 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:14:51.567482 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 28 00:14:51.609662 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 28 00:14:51.991175 systemd-networkd[1371]: eth1: Gained IPv6LL Apr 28 00:14:51.992907 systemd-timesyncd[1341]: Network configuration changed, trying to establish connection. Apr 28 00:14:52.342572 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:14:52.345302 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 28 00:14:52.351041 systemd[1]: Startup finished in 810ms (kernel) + 5.211s (initrd) + 4.480s (userspace) = 10.503s. Apr 28 00:14:52.354789 (kubelet)[1576]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 28 00:14:52.896639 kubelet[1576]: E0428 00:14:52.896574 1576 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 28 00:14:52.901099 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 28 00:14:52.901443 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 28 00:15:02.909542 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 28 00:15:02.919162 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:03.037793 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:03.051022 (kubelet)[1595]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 28 00:15:03.101629 kubelet[1595]: E0428 00:15:03.101534 1595 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 28 00:15:03.106804 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 28 00:15:03.107150 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 28 00:15:13.159682 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 28 00:15:13.178547 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:13.300478 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:13.306255 (kubelet)[1610]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 28 00:15:13.344622 kubelet[1610]: E0428 00:15:13.344486 1610 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 28 00:15:13.347452 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 28 00:15:13.347646 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 28 00:15:20.986897 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 28 00:15:20.996542 systemd[1]: Started sshd@0-178.105.21.241:22-50.85.169.122:42890.service - OpenSSH per-connection server daemon (50.85.169.122:42890). Apr 28 00:15:21.130121 sshd[1617]: Accepted publickey for core from 50.85.169.122 port 42890 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:15:21.134257 sshd[1617]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:15:21.147309 systemd-logind[1454]: New session 1 of user core. Apr 28 00:15:21.148963 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 28 00:15:21.155293 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 28 00:15:21.170226 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 28 00:15:21.183311 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 28 00:15:21.188829 (systemd)[1621]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 28 00:15:21.293023 systemd[1621]: Queued start job for default target default.target. Apr 28 00:15:21.302019 systemd[1621]: Created slice app.slice - User Application Slice. Apr 28 00:15:21.302057 systemd[1621]: Reached target paths.target - Paths. Apr 28 00:15:21.302072 systemd[1621]: Reached target timers.target - Timers. Apr 28 00:15:21.304122 systemd[1621]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 28 00:15:21.320243 systemd[1621]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 28 00:15:21.320393 systemd[1621]: Reached target sockets.target - Sockets. Apr 28 00:15:21.320409 systemd[1621]: Reached target basic.target - Basic System. Apr 28 00:15:21.320457 systemd[1621]: Reached target default.target - Main User Target. Apr 28 00:15:21.320510 systemd[1621]: Startup finished in 124ms. Apr 28 00:15:21.320640 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 28 00:15:21.331058 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 28 00:15:21.448637 systemd[1]: Started sshd@1-178.105.21.241:22-50.85.169.122:42900.service - OpenSSH per-connection server daemon (50.85.169.122:42900). Apr 28 00:15:21.568046 sshd[1632]: Accepted publickey for core from 50.85.169.122 port 42900 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:15:21.569898 sshd[1632]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:15:21.579020 systemd-logind[1454]: New session 2 of user core. Apr 28 00:15:21.589060 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 28 00:15:21.686780 sshd[1632]: pam_unix(sshd:session): session closed for user core Apr 28 00:15:21.695606 systemd[1]: sshd@1-178.105.21.241:22-50.85.169.122:42900.service: Deactivated successfully. Apr 28 00:15:21.699533 systemd[1]: session-2.scope: Deactivated successfully. Apr 28 00:15:21.701420 systemd-logind[1454]: Session 2 logged out. Waiting for processes to exit. Apr 28 00:15:21.710648 systemd-logind[1454]: Removed session 2. Apr 28 00:15:21.718296 systemd[1]: Started sshd@2-178.105.21.241:22-50.85.169.122:42912.service - OpenSSH per-connection server daemon (50.85.169.122:42912). Apr 28 00:15:21.842263 sshd[1639]: Accepted publickey for core from 50.85.169.122 port 42912 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:15:21.845309 sshd[1639]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:15:21.851979 systemd-logind[1454]: New session 3 of user core. Apr 28 00:15:21.861013 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 28 00:15:21.959199 sshd[1639]: pam_unix(sshd:session): session closed for user core Apr 28 00:15:21.965330 systemd[1]: sshd@2-178.105.21.241:22-50.85.169.122:42912.service: Deactivated successfully. Apr 28 00:15:21.968600 systemd[1]: session-3.scope: Deactivated successfully. Apr 28 00:15:21.970888 systemd-logind[1454]: Session 3 logged out. Waiting for processes to exit. Apr 28 00:15:21.980090 systemd-logind[1454]: Removed session 3. Apr 28 00:15:21.986152 systemd[1]: Started sshd@3-178.105.21.241:22-50.85.169.122:42924.service - OpenSSH per-connection server daemon (50.85.169.122:42924). Apr 28 00:15:22.113875 sshd[1646]: Accepted publickey for core from 50.85.169.122 port 42924 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:15:22.116649 sshd[1646]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:15:22.124984 systemd-logind[1454]: New session 4 of user core. Apr 28 00:15:22.137093 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 28 00:15:22.241345 sshd[1646]: pam_unix(sshd:session): session closed for user core Apr 28 00:15:22.246973 systemd-logind[1454]: Session 4 logged out. Waiting for processes to exit. Apr 28 00:15:22.247369 systemd[1]: sshd@3-178.105.21.241:22-50.85.169.122:42924.service: Deactivated successfully. Apr 28 00:15:22.249123 systemd[1]: session-4.scope: Deactivated successfully. Apr 28 00:15:22.250340 systemd-logind[1454]: Removed session 4. Apr 28 00:15:22.270046 systemd[1]: Started sshd@4-178.105.21.241:22-50.85.169.122:42926.service - OpenSSH per-connection server daemon (50.85.169.122:42926). Apr 28 00:15:21.885865 systemd-resolved[1326]: Clock change detected. Flushing caches. Apr 28 00:15:21.895277 systemd-journald[1115]: Time jumped backwards, rotating. Apr 28 00:15:21.886130 systemd-timesyncd[1341]: Contacted time server 162.159.200.1:123 (2.flatcar.pool.ntp.org). Apr 28 00:15:21.886236 systemd-timesyncd[1341]: Initial clock synchronization to Tue 2026-04-28 00:15:21.885718 UTC. Apr 28 00:15:21.946173 sshd[1653]: Accepted publickey for core from 50.85.169.122 port 42926 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:15:21.949033 sshd[1653]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:15:21.957626 systemd-logind[1454]: New session 5 of user core. Apr 28 00:15:21.964029 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 28 00:15:22.062421 sudo[1657]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 28 00:15:22.063131 sudo[1657]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 28 00:15:22.085894 sudo[1657]: pam_unix(sudo:session): session closed for user root Apr 28 00:15:22.104226 sshd[1653]: pam_unix(sshd:session): session closed for user core Apr 28 00:15:22.111177 systemd-logind[1454]: Session 5 logged out. Waiting for processes to exit. Apr 28 00:15:22.111376 systemd[1]: sshd@4-178.105.21.241:22-50.85.169.122:42926.service: Deactivated successfully. Apr 28 00:15:22.113480 systemd[1]: session-5.scope: Deactivated successfully. Apr 28 00:15:22.125828 systemd-logind[1454]: Removed session 5. Apr 28 00:15:22.134088 systemd[1]: Started sshd@5-178.105.21.241:22-50.85.169.122:42932.service - OpenSSH per-connection server daemon (50.85.169.122:42932). Apr 28 00:15:22.250748 sshd[1662]: Accepted publickey for core from 50.85.169.122 port 42932 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:15:22.253946 sshd[1662]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:15:22.259183 systemd-logind[1454]: New session 6 of user core. Apr 28 00:15:22.270000 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 28 00:15:22.356295 sudo[1666]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 28 00:15:22.357016 sudo[1666]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 28 00:15:22.361452 sudo[1666]: pam_unix(sudo:session): session closed for user root Apr 28 00:15:22.367562 sudo[1665]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Apr 28 00:15:22.367968 sudo[1665]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 28 00:15:22.386229 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Apr 28 00:15:22.388841 auditctl[1669]: No rules Apr 28 00:15:22.389759 systemd[1]: audit-rules.service: Deactivated successfully. Apr 28 00:15:22.389991 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Apr 28 00:15:22.392119 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 28 00:15:22.440270 augenrules[1687]: No rules Apr 28 00:15:22.442346 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 28 00:15:22.445599 sudo[1665]: pam_unix(sudo:session): session closed for user root Apr 28 00:15:22.460857 sshd[1662]: pam_unix(sshd:session): session closed for user core Apr 28 00:15:22.467162 systemd-logind[1454]: Session 6 logged out. Waiting for processes to exit. Apr 28 00:15:22.467328 systemd[1]: sshd@5-178.105.21.241:22-50.85.169.122:42932.service: Deactivated successfully. Apr 28 00:15:22.469215 systemd[1]: session-6.scope: Deactivated successfully. Apr 28 00:15:22.470499 systemd-logind[1454]: Removed session 6. Apr 28 00:15:22.486252 systemd[1]: Started sshd@6-178.105.21.241:22-50.85.169.122:42942.service - OpenSSH per-connection server daemon (50.85.169.122:42942). Apr 28 00:15:22.622371 sshd[1695]: Accepted publickey for core from 50.85.169.122 port 42942 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:15:22.625874 sshd[1695]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:15:22.631492 systemd-logind[1454]: New session 7 of user core. Apr 28 00:15:22.642002 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 28 00:15:22.727963 sudo[1698]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 28 00:15:22.728240 sudo[1698]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 28 00:15:22.958866 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Apr 28 00:15:22.970465 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:23.049097 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 28 00:15:23.050642 (dockerd)[1717]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 28 00:15:23.094846 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:23.098488 (kubelet)[1723]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 28 00:15:23.148799 kubelet[1723]: E0428 00:15:23.148656 1723 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 28 00:15:23.152338 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 28 00:15:23.152502 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 28 00:15:23.309812 dockerd[1717]: time="2026-04-28T00:15:23.309608370Z" level=info msg="Starting up" Apr 28 00:15:23.410167 dockerd[1717]: time="2026-04-28T00:15:23.409868010Z" level=info msg="Loading containers: start." Apr 28 00:15:23.514700 kernel: Initializing XFRM netlink socket Apr 28 00:15:23.602799 systemd-networkd[1371]: docker0: Link UP Apr 28 00:15:23.618486 dockerd[1717]: time="2026-04-28T00:15:23.618433370Z" level=info msg="Loading containers: done." Apr 28 00:15:23.636442 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2571548780-merged.mount: Deactivated successfully. Apr 28 00:15:23.638187 dockerd[1717]: time="2026-04-28T00:15:23.637553530Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 28 00:15:23.639102 dockerd[1717]: time="2026-04-28T00:15:23.638725610Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Apr 28 00:15:23.639102 dockerd[1717]: time="2026-04-28T00:15:23.638896850Z" level=info msg="Daemon has completed initialization" Apr 28 00:15:23.689379 dockerd[1717]: time="2026-04-28T00:15:23.689241690Z" level=info msg="API listen on /run/docker.sock" Apr 28 00:15:23.690509 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 28 00:15:24.192185 containerd[1481]: time="2026-04-28T00:15:24.191906050Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\"" Apr 28 00:15:24.780977 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3740113033.mount: Deactivated successfully. Apr 28 00:15:25.779776 containerd[1481]: time="2026-04-28T00:15:25.779239290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:25.781141 containerd[1481]: time="2026-04-28T00:15:25.781077970Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.11: active requests=0, bytes read=27008885" Apr 28 00:15:25.782219 containerd[1481]: time="2026-04-28T00:15:25.781702290Z" level=info msg="ImageCreate event name:\"sha256:51b83c5cb2f791f72696c040be904535bad3c81a6ffc19a55013ac150a24d9b0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:25.786508 containerd[1481]: time="2026-04-28T00:15:25.786457970Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:25.787990 containerd[1481]: time="2026-04-28T00:15:25.787947810Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.11\" with image id \"sha256:51b83c5cb2f791f72696c040be904535bad3c81a6ffc19a55013ac150a24d9b0\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\", size \"27005386\" in 1.59599228s" Apr 28 00:15:25.788127 containerd[1481]: time="2026-04-28T00:15:25.788106530Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\" returns image reference \"sha256:51b83c5cb2f791f72696c040be904535bad3c81a6ffc19a55013ac150a24d9b0\"" Apr 28 00:15:25.789168 containerd[1481]: time="2026-04-28T00:15:25.789093530Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\"" Apr 28 00:15:26.919077 containerd[1481]: time="2026-04-28T00:15:26.918994010Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:26.921151 containerd[1481]: time="2026-04-28T00:15:26.921053690Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.11: active requests=0, bytes read=23297794" Apr 28 00:15:26.921592 containerd[1481]: time="2026-04-28T00:15:26.921491690Z" level=info msg="ImageCreate event name:\"sha256:df8bcecad66863646fb4016494163838761da38376bae5a7592e04041db8489a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:26.925408 containerd[1481]: time="2026-04-28T00:15:26.925343170Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:26.926663 containerd[1481]: time="2026-04-28T00:15:26.926469730Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.11\" with image id \"sha256:df8bcecad66863646fb4016494163838761da38376bae5a7592e04041db8489a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\", size \"24804413\" in 1.13720804s" Apr 28 00:15:26.926663 containerd[1481]: time="2026-04-28T00:15:26.926554970Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\" returns image reference \"sha256:df8bcecad66863646fb4016494163838761da38376bae5a7592e04041db8489a\"" Apr 28 00:15:26.927610 containerd[1481]: time="2026-04-28T00:15:26.927440210Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\"" Apr 28 00:15:27.860778 containerd[1481]: time="2026-04-28T00:15:27.860690610Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:27.862981 containerd[1481]: time="2026-04-28T00:15:27.862897770Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.11: active requests=0, bytes read=18141378" Apr 28 00:15:27.865355 containerd[1481]: time="2026-04-28T00:15:27.864430690Z" level=info msg="ImageCreate event name:\"sha256:8c8e25fd00e5c108fb9ab5490c25bfaeb0231b1c59f749dab4f5300f1c49995b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:27.875268 containerd[1481]: time="2026-04-28T00:15:27.874989130Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:27.876463 containerd[1481]: time="2026-04-28T00:15:27.876260610Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.11\" with image id \"sha256:8c8e25fd00e5c108fb9ab5490c25bfaeb0231b1c59f749dab4f5300f1c49995b\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\", size \"19648015\" in 948.77504ms" Apr 28 00:15:27.876463 containerd[1481]: time="2026-04-28T00:15:27.876301050Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\" returns image reference \"sha256:8c8e25fd00e5c108fb9ab5490c25bfaeb0231b1c59f749dab4f5300f1c49995b\"" Apr 28 00:15:27.877240 containerd[1481]: time="2026-04-28T00:15:27.877193370Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\"" Apr 28 00:15:28.770859 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2664212939.mount: Deactivated successfully. Apr 28 00:15:29.138788 containerd[1481]: time="2026-04-28T00:15:29.138283530Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:29.140734 containerd[1481]: time="2026-04-28T00:15:29.140658210Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.11: active requests=0, bytes read=28040534" Apr 28 00:15:29.144160 containerd[1481]: time="2026-04-28T00:15:29.143822170Z" level=info msg="ImageCreate event name:\"sha256:7ce14d6fb1e5134a578d2aaa327fd701273e3d222b9b8d88054dd86b87a7dc36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:29.149103 containerd[1481]: time="2026-04-28T00:15:29.149023810Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:29.150450 containerd[1481]: time="2026-04-28T00:15:29.150388130Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.11\" with image id \"sha256:7ce14d6fb1e5134a578d2aaa327fd701273e3d222b9b8d88054dd86b87a7dc36\", repo tag \"registry.k8s.io/kube-proxy:v1.33.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\", size \"28039527\" in 1.27301544s" Apr 28 00:15:29.150951 containerd[1481]: time="2026-04-28T00:15:29.150719610Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\" returns image reference \"sha256:7ce14d6fb1e5134a578d2aaa327fd701273e3d222b9b8d88054dd86b87a7dc36\"" Apr 28 00:15:29.151647 containerd[1481]: time="2026-04-28T00:15:29.151518010Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Apr 28 00:15:29.748383 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3271070095.mount: Deactivated successfully. Apr 28 00:15:30.465093 containerd[1481]: time="2026-04-28T00:15:30.464588250Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:30.466364 containerd[1481]: time="2026-04-28T00:15:30.466312570Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=19152209" Apr 28 00:15:30.470732 containerd[1481]: time="2026-04-28T00:15:30.469722130Z" level=info msg="ImageCreate event name:\"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:30.472634 containerd[1481]: time="2026-04-28T00:15:30.472577530Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:30.475506 containerd[1481]: time="2026-04-28T00:15:30.475446850Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"19148915\" in 1.32363828s" Apr 28 00:15:30.475709 containerd[1481]: time="2026-04-28T00:15:30.475690730Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\"" Apr 28 00:15:30.476332 containerd[1481]: time="2026-04-28T00:15:30.476240410Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 28 00:15:30.924525 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount173755408.mount: Deactivated successfully. Apr 28 00:15:30.935651 containerd[1481]: time="2026-04-28T00:15:30.935486650Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:30.938405 containerd[1481]: time="2026-04-28T00:15:30.938136050Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268723" Apr 28 00:15:30.939545 containerd[1481]: time="2026-04-28T00:15:30.939462570Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:30.942929 containerd[1481]: time="2026-04-28T00:15:30.942846690Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:30.945753 containerd[1481]: time="2026-04-28T00:15:30.943888970Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 467.28608ms" Apr 28 00:15:30.945753 containerd[1481]: time="2026-04-28T00:15:30.943942970Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Apr 28 00:15:30.946395 containerd[1481]: time="2026-04-28T00:15:30.946347650Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Apr 28 00:15:31.462131 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1216486372.mount: Deactivated successfully. Apr 28 00:15:32.315735 containerd[1481]: time="2026-04-28T00:15:32.315014410Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:32.316314 containerd[1481]: time="2026-04-28T00:15:32.316238610Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=21886470" Apr 28 00:15:32.318699 containerd[1481]: time="2026-04-28T00:15:32.318183210Z" level=info msg="ImageCreate event name:\"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:32.323745 containerd[1481]: time="2026-04-28T00:15:32.323645170Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:32.325351 containerd[1481]: time="2026-04-28T00:15:32.325053450Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"21882972\" in 1.37828976s" Apr 28 00:15:32.325351 containerd[1481]: time="2026-04-28T00:15:32.325110650Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\"" Apr 28 00:15:33.207960 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Apr 28 00:15:33.216964 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:33.367983 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:33.370621 (kubelet)[2089]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 28 00:15:33.420707 kubelet[2089]: E0428 00:15:33.419861 2089 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 28 00:15:33.423444 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 28 00:15:33.423610 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 28 00:15:35.408813 update_engine[1455]: I20260428 00:15:35.408704 1455 update_attempter.cc:509] Updating boot flags... Apr 28 00:15:35.467762 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (2105) Apr 28 00:15:35.544919 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (2104) Apr 28 00:15:38.077055 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:38.087060 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:38.122979 systemd[1]: Reloading requested from client PID 2121 ('systemctl') (unit session-7.scope)... Apr 28 00:15:38.122994 systemd[1]: Reloading... Apr 28 00:15:38.239726 zram_generator::config[2158]: No configuration found. Apr 28 00:15:38.374481 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 28 00:15:38.445605 systemd[1]: Reloading finished in 322 ms. Apr 28 00:15:38.498664 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 28 00:15:38.498944 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 28 00:15:38.500784 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:38.513303 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:38.639511 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:38.654703 (kubelet)[2209]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 28 00:15:38.699198 kubelet[2209]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 28 00:15:38.699198 kubelet[2209]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 28 00:15:38.699198 kubelet[2209]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 28 00:15:38.699716 kubelet[2209]: I0428 00:15:38.699273 2209 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 28 00:15:39.248111 kubelet[2209]: I0428 00:15:39.248044 2209 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 28 00:15:39.248111 kubelet[2209]: I0428 00:15:39.248091 2209 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 28 00:15:39.248446 kubelet[2209]: I0428 00:15:39.248405 2209 server.go:956] "Client rotation is on, will bootstrap in background" Apr 28 00:15:39.281314 kubelet[2209]: E0428 00:15:39.281256 2209 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://178.105.21.241:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 178.105.21.241:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 28 00:15:39.281764 kubelet[2209]: I0428 00:15:39.281595 2209 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 28 00:15:39.295557 kubelet[2209]: E0428 00:15:39.295479 2209 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 28 00:15:39.296706 kubelet[2209]: I0428 00:15:39.295795 2209 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 28 00:15:39.300756 kubelet[2209]: I0428 00:15:39.300722 2209 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 28 00:15:39.304317 kubelet[2209]: I0428 00:15:39.304245 2209 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 28 00:15:39.304825 kubelet[2209]: I0428 00:15:39.304496 2209 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-7-n-51c70c830d","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 28 00:15:39.305025 kubelet[2209]: I0428 00:15:39.305007 2209 topology_manager.go:138] "Creating topology manager with none policy" Apr 28 00:15:39.305111 kubelet[2209]: I0428 00:15:39.305099 2209 container_manager_linux.go:303] "Creating device plugin manager" Apr 28 00:15:39.305454 kubelet[2209]: I0428 00:15:39.305436 2209 state_mem.go:36] "Initialized new in-memory state store" Apr 28 00:15:39.309378 kubelet[2209]: I0428 00:15:39.309341 2209 kubelet.go:480] "Attempting to sync node with API server" Apr 28 00:15:39.309574 kubelet[2209]: I0428 00:15:39.309559 2209 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 28 00:15:39.309640 kubelet[2209]: I0428 00:15:39.309632 2209 kubelet.go:386] "Adding apiserver pod source" Apr 28 00:15:39.311255 kubelet[2209]: I0428 00:15:39.311226 2209 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 28 00:15:39.315489 kubelet[2209]: E0428 00:15:39.315427 2209 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://178.105.21.241:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-7-n-51c70c830d&limit=500&resourceVersion=0\": dial tcp 178.105.21.241:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 28 00:15:39.316189 kubelet[2209]: E0428 00:15:39.316148 2209 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://178.105.21.241:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 178.105.21.241:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 28 00:15:39.316650 kubelet[2209]: I0428 00:15:39.316626 2209 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 28 00:15:39.317523 kubelet[2209]: I0428 00:15:39.317493 2209 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 28 00:15:39.317662 kubelet[2209]: W0428 00:15:39.317647 2209 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 28 00:15:39.323732 kubelet[2209]: I0428 00:15:39.322889 2209 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 28 00:15:39.323732 kubelet[2209]: I0428 00:15:39.322951 2209 server.go:1289] "Started kubelet" Apr 28 00:15:39.324887 kubelet[2209]: I0428 00:15:39.324835 2209 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 28 00:15:39.331993 kubelet[2209]: I0428 00:15:39.331929 2209 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 28 00:15:39.333214 kubelet[2209]: I0428 00:15:39.333170 2209 server.go:317] "Adding debug handlers to kubelet server" Apr 28 00:15:39.336167 kubelet[2209]: I0428 00:15:39.336119 2209 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 28 00:15:39.336706 kubelet[2209]: E0428 00:15:39.336436 2209 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-7-n-51c70c830d\" not found" Apr 28 00:15:39.337889 kubelet[2209]: I0428 00:15:39.337352 2209 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 28 00:15:39.337889 kubelet[2209]: I0428 00:15:39.337597 2209 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 28 00:15:39.340862 kubelet[2209]: I0428 00:15:39.340815 2209 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 28 00:15:39.340982 kubelet[2209]: I0428 00:15:39.340907 2209 reconciler.go:26] "Reconciler: start to sync state" Apr 28 00:15:39.341193 kubelet[2209]: I0428 00:15:39.341173 2209 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 28 00:15:39.345667 kubelet[2209]: E0428 00:15:39.343419 2209 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://178.105.21.241:6443/api/v1/namespaces/default/events\": dial tcp 178.105.21.241:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081-3-7-n-51c70c830d.18aa5d130b17415a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-3-7-n-51c70c830d,UID:ci-4081-3-7-n-51c70c830d,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-3-7-n-51c70c830d,},FirstTimestamp:2026-04-28 00:15:39.32291721 +0000 UTC m=+0.661740921,LastTimestamp:2026-04-28 00:15:39.32291721 +0000 UTC m=+0.661740921,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-7-n-51c70c830d,}" Apr 28 00:15:39.346349 kubelet[2209]: I0428 00:15:39.346314 2209 factory.go:223] Registration of the systemd container factory successfully Apr 28 00:15:39.346566 kubelet[2209]: I0428 00:15:39.346542 2209 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 28 00:15:39.348096 kubelet[2209]: E0428 00:15:39.348057 2209 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://178.105.21.241:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-7-n-51c70c830d?timeout=10s\": dial tcp 178.105.21.241:6443: connect: connection refused" interval="200ms" Apr 28 00:15:39.349850 kubelet[2209]: E0428 00:15:39.349244 2209 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 28 00:15:39.353725 kubelet[2209]: I0428 00:15:39.352487 2209 factory.go:223] Registration of the containerd container factory successfully Apr 28 00:15:39.354397 kubelet[2209]: I0428 00:15:39.354336 2209 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 28 00:15:39.358353 kubelet[2209]: I0428 00:15:39.358303 2209 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 28 00:15:39.358353 kubelet[2209]: I0428 00:15:39.358352 2209 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 28 00:15:39.358503 kubelet[2209]: I0428 00:15:39.358393 2209 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 28 00:15:39.358503 kubelet[2209]: I0428 00:15:39.358406 2209 kubelet.go:2436] "Starting kubelet main sync loop" Apr 28 00:15:39.358503 kubelet[2209]: E0428 00:15:39.358469 2209 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 28 00:15:39.369918 kubelet[2209]: E0428 00:15:39.369839 2209 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://178.105.21.241:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 178.105.21.241:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 28 00:15:39.375318 kubelet[2209]: E0428 00:15:39.375246 2209 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://178.105.21.241:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 178.105.21.241:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 28 00:15:39.384694 kubelet[2209]: I0428 00:15:39.384399 2209 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 28 00:15:39.384694 kubelet[2209]: I0428 00:15:39.384419 2209 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 28 00:15:39.384694 kubelet[2209]: I0428 00:15:39.384439 2209 state_mem.go:36] "Initialized new in-memory state store" Apr 28 00:15:39.386953 kubelet[2209]: I0428 00:15:39.386922 2209 policy_none.go:49] "None policy: Start" Apr 28 00:15:39.386953 kubelet[2209]: I0428 00:15:39.386957 2209 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 28 00:15:39.387073 kubelet[2209]: I0428 00:15:39.386972 2209 state_mem.go:35] "Initializing new in-memory state store" Apr 28 00:15:39.395412 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 28 00:15:39.413713 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 28 00:15:39.417830 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 28 00:15:39.426774 kubelet[2209]: E0428 00:15:39.426725 2209 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 28 00:15:39.428545 kubelet[2209]: I0428 00:15:39.427613 2209 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 28 00:15:39.428545 kubelet[2209]: I0428 00:15:39.427646 2209 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 28 00:15:39.428545 kubelet[2209]: I0428 00:15:39.428333 2209 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 28 00:15:39.430614 kubelet[2209]: E0428 00:15:39.430085 2209 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 28 00:15:39.430614 kubelet[2209]: E0428 00:15:39.430131 2209 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081-3-7-n-51c70c830d\" not found" Apr 28 00:15:39.475699 systemd[1]: Created slice kubepods-burstable-pod4c3f259df7978e32694e0430f8ef34c0.slice - libcontainer container kubepods-burstable-pod4c3f259df7978e32694e0430f8ef34c0.slice. Apr 28 00:15:39.486608 kubelet[2209]: E0428 00:15:39.485619 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.490634 systemd[1]: Created slice kubepods-burstable-podaacf343c9c5f354d3536c75cc0d11c90.slice - libcontainer container kubepods-burstable-podaacf343c9c5f354d3536c75cc0d11c90.slice. Apr 28 00:15:39.495610 kubelet[2209]: E0428 00:15:39.495576 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.497509 systemd[1]: Created slice kubepods-burstable-pode5ef54897d31f1c16dbf3b1d0bbbfab2.slice - libcontainer container kubepods-burstable-pode5ef54897d31f1c16dbf3b1d0bbbfab2.slice. Apr 28 00:15:39.499651 kubelet[2209]: E0428 00:15:39.499531 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.531476 kubelet[2209]: I0428 00:15:39.531428 2209 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.532033 kubelet[2209]: E0428 00:15:39.531978 2209 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://178.105.21.241:6443/api/v1/nodes\": dial tcp 178.105.21.241:6443: connect: connection refused" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.549441 kubelet[2209]: E0428 00:15:39.549365 2209 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://178.105.21.241:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-7-n-51c70c830d?timeout=10s\": dial tcp 178.105.21.241:6443: connect: connection refused" interval="400ms" Apr 28 00:15:39.641983 kubelet[2209]: I0428 00:15:39.641901 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.641983 kubelet[2209]: I0428 00:15:39.641970 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4c3f259df7978e32694e0430f8ef34c0-kubeconfig\") pod \"kube-scheduler-ci-4081-3-7-n-51c70c830d\" (UID: \"4c3f259df7978e32694e0430f8ef34c0\") " pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.642233 kubelet[2209]: I0428 00:15:39.642005 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aacf343c9c5f354d3536c75cc0d11c90-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" (UID: \"aacf343c9c5f354d3536c75cc0d11c90\") " pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.642233 kubelet[2209]: I0428 00:15:39.642062 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.642233 kubelet[2209]: I0428 00:15:39.642093 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.642233 kubelet[2209]: I0428 00:15:39.642121 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aacf343c9c5f354d3536c75cc0d11c90-ca-certs\") pod \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" (UID: \"aacf343c9c5f354d3536c75cc0d11c90\") " pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.642233 kubelet[2209]: I0428 00:15:39.642146 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aacf343c9c5f354d3536c75cc0d11c90-k8s-certs\") pod \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" (UID: \"aacf343c9c5f354d3536c75cc0d11c90\") " pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.642470 kubelet[2209]: I0428 00:15:39.642177 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-ca-certs\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.642470 kubelet[2209]: I0428 00:15:39.642218 2209 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.735003 kubelet[2209]: I0428 00:15:39.734919 2209 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.735579 kubelet[2209]: E0428 00:15:39.735479 2209 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://178.105.21.241:6443/api/v1/nodes\": dial tcp 178.105.21.241:6443: connect: connection refused" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:39.789852 containerd[1481]: time="2026-04-28T00:15:39.789188730Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-7-n-51c70c830d,Uid:4c3f259df7978e32694e0430f8ef34c0,Namespace:kube-system,Attempt:0,}" Apr 28 00:15:39.797621 containerd[1481]: time="2026-04-28T00:15:39.797559890Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-7-n-51c70c830d,Uid:aacf343c9c5f354d3536c75cc0d11c90,Namespace:kube-system,Attempt:0,}" Apr 28 00:15:39.802772 containerd[1481]: time="2026-04-28T00:15:39.802376890Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-7-n-51c70c830d,Uid:e5ef54897d31f1c16dbf3b1d0bbbfab2,Namespace:kube-system,Attempt:0,}" Apr 28 00:15:39.951471 kubelet[2209]: E0428 00:15:39.951387 2209 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://178.105.21.241:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-7-n-51c70c830d?timeout=10s\": dial tcp 178.105.21.241:6443: connect: connection refused" interval="800ms" Apr 28 00:15:40.138501 kubelet[2209]: I0428 00:15:40.137940 2209 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:40.138501 kubelet[2209]: E0428 00:15:40.138355 2209 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://178.105.21.241:6443/api/v1/nodes\": dial tcp 178.105.21.241:6443: connect: connection refused" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:40.179031 kubelet[2209]: E0428 00:15:40.178939 2209 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://178.105.21.241:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 178.105.21.241:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 28 00:15:40.274729 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2775744572.mount: Deactivated successfully. Apr 28 00:15:40.285272 containerd[1481]: time="2026-04-28T00:15:40.285138770Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 28 00:15:40.288542 containerd[1481]: time="2026-04-28T00:15:40.288436090Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 28 00:15:40.289898 containerd[1481]: time="2026-04-28T00:15:40.289779130Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269193" Apr 28 00:15:40.291568 containerd[1481]: time="2026-04-28T00:15:40.291518890Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 28 00:15:40.293736 containerd[1481]: time="2026-04-28T00:15:40.293654890Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 28 00:15:40.297259 containerd[1481]: time="2026-04-28T00:15:40.296379130Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 28 00:15:40.297259 containerd[1481]: time="2026-04-28T00:15:40.296521890Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 28 00:15:40.300781 containerd[1481]: time="2026-04-28T00:15:40.300713930Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 28 00:15:40.304699 containerd[1481]: time="2026-04-28T00:15:40.303302410Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 505.5946ms" Apr 28 00:15:40.305178 containerd[1481]: time="2026-04-28T00:15:40.305145690Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 515.80248ms" Apr 28 00:15:40.305986 containerd[1481]: time="2026-04-28T00:15:40.305957290Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 503.49936ms" Apr 28 00:15:40.439353 containerd[1481]: time="2026-04-28T00:15:40.438940810Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:15:40.439353 containerd[1481]: time="2026-04-28T00:15:40.439022490Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:15:40.439353 containerd[1481]: time="2026-04-28T00:15:40.439035810Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:40.439353 containerd[1481]: time="2026-04-28T00:15:40.439170410Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:40.441995 containerd[1481]: time="2026-04-28T00:15:40.441844130Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:15:40.442965 containerd[1481]: time="2026-04-28T00:15:40.442884530Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:15:40.443101 containerd[1481]: time="2026-04-28T00:15:40.442987970Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:40.443221 containerd[1481]: time="2026-04-28T00:15:40.443174730Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:40.448494 containerd[1481]: time="2026-04-28T00:15:40.448028210Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:15:40.448494 containerd[1481]: time="2026-04-28T00:15:40.448116130Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:15:40.448494 containerd[1481]: time="2026-04-28T00:15:40.448130810Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:40.448494 containerd[1481]: time="2026-04-28T00:15:40.448299650Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:40.475141 systemd[1]: Started cri-containerd-b758e01b7fed5ba3cfafabab26c0cf47a984f868c9b30445f29b534230d8be15.scope - libcontainer container b758e01b7fed5ba3cfafabab26c0cf47a984f868c9b30445f29b534230d8be15. Apr 28 00:15:40.480773 systemd[1]: Started cri-containerd-d54450c6df20b4b98a4550cf7e0b028ff4fa4297a3b18073e11264327f670077.scope - libcontainer container d54450c6df20b4b98a4550cf7e0b028ff4fa4297a3b18073e11264327f670077. Apr 28 00:15:40.491882 systemd[1]: Started cri-containerd-b0b00487ccd2a4836a1069a46dc7676145c6687188beb0f0ce7e427398d4fc3a.scope - libcontainer container b0b00487ccd2a4836a1069a46dc7676145c6687188beb0f0ce7e427398d4fc3a. Apr 28 00:15:40.537887 containerd[1481]: time="2026-04-28T00:15:40.537794330Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-7-n-51c70c830d,Uid:aacf343c9c5f354d3536c75cc0d11c90,Namespace:kube-system,Attempt:0,} returns sandbox id \"b0b00487ccd2a4836a1069a46dc7676145c6687188beb0f0ce7e427398d4fc3a\"" Apr 28 00:15:40.544453 containerd[1481]: time="2026-04-28T00:15:40.544343690Z" level=info msg="CreateContainer within sandbox \"b0b00487ccd2a4836a1069a46dc7676145c6687188beb0f0ce7e427398d4fc3a\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 28 00:15:40.551463 containerd[1481]: time="2026-04-28T00:15:40.551174650Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-7-n-51c70c830d,Uid:4c3f259df7978e32694e0430f8ef34c0,Namespace:kube-system,Attempt:0,} returns sandbox id \"d54450c6df20b4b98a4550cf7e0b028ff4fa4297a3b18073e11264327f670077\"" Apr 28 00:15:40.558108 containerd[1481]: time="2026-04-28T00:15:40.558065010Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-7-n-51c70c830d,Uid:e5ef54897d31f1c16dbf3b1d0bbbfab2,Namespace:kube-system,Attempt:0,} returns sandbox id \"b758e01b7fed5ba3cfafabab26c0cf47a984f868c9b30445f29b534230d8be15\"" Apr 28 00:15:40.560734 containerd[1481]: time="2026-04-28T00:15:40.560618770Z" level=info msg="CreateContainer within sandbox \"d54450c6df20b4b98a4550cf7e0b028ff4fa4297a3b18073e11264327f670077\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 28 00:15:40.563134 containerd[1481]: time="2026-04-28T00:15:40.563024970Z" level=info msg="CreateContainer within sandbox \"b758e01b7fed5ba3cfafabab26c0cf47a984f868c9b30445f29b534230d8be15\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 28 00:15:40.574827 containerd[1481]: time="2026-04-28T00:15:40.574717330Z" level=info msg="CreateContainer within sandbox \"b0b00487ccd2a4836a1069a46dc7676145c6687188beb0f0ce7e427398d4fc3a\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0338f6f7b29563117958a767f7ae4d0405ee0bf630353bd937b14005e02946a1\"" Apr 28 00:15:40.577585 containerd[1481]: time="2026-04-28T00:15:40.576047730Z" level=info msg="StartContainer for \"0338f6f7b29563117958a767f7ae4d0405ee0bf630353bd937b14005e02946a1\"" Apr 28 00:15:40.579713 kubelet[2209]: E0428 00:15:40.579625 2209 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://178.105.21.241:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 178.105.21.241:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 28 00:15:40.589826 containerd[1481]: time="2026-04-28T00:15:40.589776010Z" level=info msg="CreateContainer within sandbox \"d54450c6df20b4b98a4550cf7e0b028ff4fa4297a3b18073e11264327f670077\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000\"" Apr 28 00:15:40.591441 containerd[1481]: time="2026-04-28T00:15:40.591402010Z" level=info msg="StartContainer for \"71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000\"" Apr 28 00:15:40.599848 containerd[1481]: time="2026-04-28T00:15:40.599499850Z" level=info msg="CreateContainer within sandbox \"b758e01b7fed5ba3cfafabab26c0cf47a984f868c9b30445f29b534230d8be15\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5\"" Apr 28 00:15:40.600874 containerd[1481]: time="2026-04-28T00:15:40.600838730Z" level=info msg="StartContainer for \"216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5\"" Apr 28 00:15:40.618211 systemd[1]: Started cri-containerd-0338f6f7b29563117958a767f7ae4d0405ee0bf630353bd937b14005e02946a1.scope - libcontainer container 0338f6f7b29563117958a767f7ae4d0405ee0bf630353bd937b14005e02946a1. Apr 28 00:15:40.645706 systemd[1]: Started cri-containerd-71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000.scope - libcontainer container 71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000. Apr 28 00:15:40.665134 systemd[1]: Started cri-containerd-216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5.scope - libcontainer container 216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5. Apr 28 00:15:40.704625 containerd[1481]: time="2026-04-28T00:15:40.704300850Z" level=info msg="StartContainer for \"0338f6f7b29563117958a767f7ae4d0405ee0bf630353bd937b14005e02946a1\" returns successfully" Apr 28 00:15:40.720836 containerd[1481]: time="2026-04-28T00:15:40.720697690Z" level=info msg="StartContainer for \"71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000\" returns successfully" Apr 28 00:15:40.743016 containerd[1481]: time="2026-04-28T00:15:40.742925090Z" level=info msg="StartContainer for \"216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5\" returns successfully" Apr 28 00:15:40.752741 kubelet[2209]: E0428 00:15:40.752635 2209 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://178.105.21.241:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-7-n-51c70c830d?timeout=10s\": dial tcp 178.105.21.241:6443: connect: connection refused" interval="1.6s" Apr 28 00:15:40.941728 kubelet[2209]: I0428 00:15:40.940655 2209 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:41.385487 kubelet[2209]: E0428 00:15:41.385124 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:41.389389 kubelet[2209]: E0428 00:15:41.389361 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:41.391710 kubelet[2209]: E0428 00:15:41.391686 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.394226 kubelet[2209]: E0428 00:15:42.393629 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.394226 kubelet[2209]: E0428 00:15:42.394029 2209 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-7-n-51c70c830d\" not found" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.657109 kubelet[2209]: I0428 00:15:42.656652 2209 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.657505 kubelet[2209]: E0428 00:15:42.657487 2209 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4081-3-7-n-51c70c830d\": node \"ci-4081-3-7-n-51c70c830d\" not found" Apr 28 00:15:42.700759 kubelet[2209]: E0428 00:15:42.700629 2209 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4081-3-7-n-51c70c830d.18aa5d130b17415a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-3-7-n-51c70c830d,UID:ci-4081-3-7-n-51c70c830d,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-3-7-n-51c70c830d,},FirstTimestamp:2026-04-28 00:15:39.32291721 +0000 UTC m=+0.661740921,LastTimestamp:2026-04-28 00:15:39.32291721 +0000 UTC m=+0.661740921,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-7-n-51c70c830d,}" Apr 28 00:15:42.718012 kubelet[2209]: E0428 00:15:42.717868 2209 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-node-lease\" not found" interval="3.2s" Apr 28 00:15:42.737290 kubelet[2209]: I0428 00:15:42.737242 2209 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.755348 kubelet[2209]: E0428 00:15:42.755049 2209 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-7-n-51c70c830d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.755348 kubelet[2209]: I0428 00:15:42.755106 2209 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.758375 kubelet[2209]: E0428 00:15:42.758119 2209 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.758375 kubelet[2209]: I0428 00:15:42.758154 2209 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:42.761331 kubelet[2209]: E0428 00:15:42.761295 2209 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:43.319139 kubelet[2209]: I0428 00:15:43.318858 2209 apiserver.go:52] "Watching apiserver" Apr 28 00:15:43.341794 kubelet[2209]: I0428 00:15:43.341736 2209 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 28 00:15:45.113408 systemd[1]: Reloading requested from client PID 2497 ('systemctl') (unit session-7.scope)... Apr 28 00:15:45.113806 systemd[1]: Reloading... Apr 28 00:15:45.223739 zram_generator::config[2538]: No configuration found. Apr 28 00:15:45.329474 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 28 00:15:45.429372 systemd[1]: Reloading finished in 315 ms. Apr 28 00:15:45.472839 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:45.489667 systemd[1]: kubelet.service: Deactivated successfully. Apr 28 00:15:45.490261 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:45.490329 systemd[1]: kubelet.service: Consumed 1.151s CPU time, 126.8M memory peak, 0B memory swap peak. Apr 28 00:15:45.499844 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 28 00:15:45.637020 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 28 00:15:45.648575 (kubelet)[2581]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 28 00:15:45.692381 kubelet[2581]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 28 00:15:45.693007 kubelet[2581]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 28 00:15:45.693195 kubelet[2581]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 28 00:15:45.693541 kubelet[2581]: I0428 00:15:45.693503 2581 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 28 00:15:45.702642 kubelet[2581]: I0428 00:15:45.702577 2581 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 28 00:15:45.702642 kubelet[2581]: I0428 00:15:45.702612 2581 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 28 00:15:45.702937 kubelet[2581]: I0428 00:15:45.702902 2581 server.go:956] "Client rotation is on, will bootstrap in background" Apr 28 00:15:45.704459 kubelet[2581]: I0428 00:15:45.704416 2581 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 28 00:15:45.707458 kubelet[2581]: I0428 00:15:45.707263 2581 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 28 00:15:45.715372 kubelet[2581]: E0428 00:15:45.715335 2581 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 28 00:15:45.715756 kubelet[2581]: I0428 00:15:45.715525 2581 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 28 00:15:45.719935 kubelet[2581]: I0428 00:15:45.719655 2581 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 28 00:15:45.721465 kubelet[2581]: I0428 00:15:45.721382 2581 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 28 00:15:45.721876 kubelet[2581]: I0428 00:15:45.721446 2581 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-7-n-51c70c830d","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 28 00:15:45.721992 kubelet[2581]: I0428 00:15:45.721914 2581 topology_manager.go:138] "Creating topology manager with none policy" Apr 28 00:15:45.721992 kubelet[2581]: I0428 00:15:45.721958 2581 container_manager_linux.go:303] "Creating device plugin manager" Apr 28 00:15:45.722169 kubelet[2581]: I0428 00:15:45.722122 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 28 00:15:45.722664 kubelet[2581]: I0428 00:15:45.722627 2581 kubelet.go:480] "Attempting to sync node with API server" Apr 28 00:15:45.722756 kubelet[2581]: I0428 00:15:45.722700 2581 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 28 00:15:45.723772 kubelet[2581]: I0428 00:15:45.722841 2581 kubelet.go:386] "Adding apiserver pod source" Apr 28 00:15:45.723772 kubelet[2581]: I0428 00:15:45.722924 2581 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 28 00:15:45.733262 kubelet[2581]: I0428 00:15:45.733230 2581 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 28 00:15:45.736018 kubelet[2581]: I0428 00:15:45.735987 2581 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 28 00:15:45.743937 kubelet[2581]: I0428 00:15:45.743909 2581 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 28 00:15:45.744421 kubelet[2581]: I0428 00:15:45.743961 2581 server.go:1289] "Started kubelet" Apr 28 00:15:45.749707 kubelet[2581]: I0428 00:15:45.748881 2581 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 28 00:15:45.760562 kubelet[2581]: I0428 00:15:45.760495 2581 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 28 00:15:45.761787 kubelet[2581]: I0428 00:15:45.761763 2581 server.go:317] "Adding debug handlers to kubelet server" Apr 28 00:15:45.765488 kubelet[2581]: I0428 00:15:45.765382 2581 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 28 00:15:45.765821 kubelet[2581]: I0428 00:15:45.765799 2581 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 28 00:15:45.766710 kubelet[2581]: I0428 00:15:45.766054 2581 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 28 00:15:45.768577 kubelet[2581]: I0428 00:15:45.768072 2581 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 28 00:15:45.768577 kubelet[2581]: E0428 00:15:45.768466 2581 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-7-n-51c70c830d\" not found" Apr 28 00:15:45.771739 kubelet[2581]: I0428 00:15:45.770490 2581 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 28 00:15:45.771739 kubelet[2581]: I0428 00:15:45.770640 2581 reconciler.go:26] "Reconciler: start to sync state" Apr 28 00:15:45.775109 kubelet[2581]: I0428 00:15:45.775058 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 28 00:15:45.777788 kubelet[2581]: I0428 00:15:45.777390 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 28 00:15:45.777788 kubelet[2581]: I0428 00:15:45.777432 2581 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 28 00:15:45.777788 kubelet[2581]: I0428 00:15:45.777453 2581 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 28 00:15:45.777788 kubelet[2581]: I0428 00:15:45.777459 2581 kubelet.go:2436] "Starting kubelet main sync loop" Apr 28 00:15:45.777788 kubelet[2581]: I0428 00:15:45.777503 2581 factory.go:223] Registration of the systemd container factory successfully Apr 28 00:15:45.777788 kubelet[2581]: E0428 00:15:45.777510 2581 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 28 00:15:45.779394 kubelet[2581]: I0428 00:15:45.778374 2581 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 28 00:15:45.783998 kubelet[2581]: I0428 00:15:45.783970 2581 factory.go:223] Registration of the containerd container factory successfully Apr 28 00:15:45.786251 kubelet[2581]: E0428 00:15:45.786204 2581 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 28 00:15:45.843236 kubelet[2581]: I0428 00:15:45.843213 2581 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 28 00:15:45.843412 kubelet[2581]: I0428 00:15:45.843397 2581 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 28 00:15:45.843474 kubelet[2581]: I0428 00:15:45.843467 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 28 00:15:45.843718 kubelet[2581]: I0428 00:15:45.843700 2581 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 28 00:15:45.844011 kubelet[2581]: I0428 00:15:45.843790 2581 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 28 00:15:45.844011 kubelet[2581]: I0428 00:15:45.843817 2581 policy_none.go:49] "None policy: Start" Apr 28 00:15:45.844011 kubelet[2581]: I0428 00:15:45.843829 2581 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 28 00:15:45.844011 kubelet[2581]: I0428 00:15:45.843843 2581 state_mem.go:35] "Initializing new in-memory state store" Apr 28 00:15:45.844011 kubelet[2581]: I0428 00:15:45.843937 2581 state_mem.go:75] "Updated machine memory state" Apr 28 00:15:45.848621 kubelet[2581]: E0428 00:15:45.848594 2581 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 28 00:15:45.849378 kubelet[2581]: I0428 00:15:45.848930 2581 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 28 00:15:45.849378 kubelet[2581]: I0428 00:15:45.848948 2581 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 28 00:15:45.849378 kubelet[2581]: I0428 00:15:45.849228 2581 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 28 00:15:45.852501 kubelet[2581]: E0428 00:15:45.850865 2581 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 28 00:15:45.878912 kubelet[2581]: I0428 00:15:45.878840 2581 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:45.879230 kubelet[2581]: I0428 00:15:45.879191 2581 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:45.879992 kubelet[2581]: I0428 00:15:45.879947 2581 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:45.953075 kubelet[2581]: I0428 00:15:45.952957 2581 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:45.974341 kubelet[2581]: I0428 00:15:45.974088 2581 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:45.974341 kubelet[2581]: I0428 00:15:45.974208 2581 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072442 kubelet[2581]: I0428 00:15:46.072172 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aacf343c9c5f354d3536c75cc0d11c90-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" (UID: \"aacf343c9c5f354d3536c75cc0d11c90\") " pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072442 kubelet[2581]: I0428 00:15:46.072220 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072442 kubelet[2581]: I0428 00:15:46.072245 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072442 kubelet[2581]: I0428 00:15:46.072265 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072442 kubelet[2581]: I0428 00:15:46.072293 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-ca-certs\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072723 kubelet[2581]: I0428 00:15:46.072324 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e5ef54897d31f1c16dbf3b1d0bbbfab2-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-7-n-51c70c830d\" (UID: \"e5ef54897d31f1c16dbf3b1d0bbbfab2\") " pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072723 kubelet[2581]: I0428 00:15:46.072343 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4c3f259df7978e32694e0430f8ef34c0-kubeconfig\") pod \"kube-scheduler-ci-4081-3-7-n-51c70c830d\" (UID: \"4c3f259df7978e32694e0430f8ef34c0\") " pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072723 kubelet[2581]: I0428 00:15:46.072359 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aacf343c9c5f354d3536c75cc0d11c90-ca-certs\") pod \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" (UID: \"aacf343c9c5f354d3536c75cc0d11c90\") " pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.072723 kubelet[2581]: I0428 00:15:46.072384 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aacf343c9c5f354d3536c75cc0d11c90-k8s-certs\") pod \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" (UID: \"aacf343c9c5f354d3536c75cc0d11c90\") " pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.723562 kubelet[2581]: I0428 00:15:46.723524 2581 apiserver.go:52] "Watching apiserver" Apr 28 00:15:46.771158 kubelet[2581]: I0428 00:15:46.771058 2581 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 28 00:15:46.823391 kubelet[2581]: I0428 00:15:46.823185 2581 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.823922 kubelet[2581]: I0428 00:15:46.823873 2581 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.832807 kubelet[2581]: E0428 00:15:46.831656 2581 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-7-n-51c70c830d\" already exists" pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.835014 kubelet[2581]: I0428 00:15:46.834858 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081-3-7-n-51c70c830d" podStartSLOduration=1.8346612900000001 podStartE2EDuration="1.83466129s" podCreationTimestamp="2026-04-28 00:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-28 00:15:46.83128097 +0000 UTC m=+1.177375201" watchObservedRunningTime="2026-04-28 00:15:46.83466129 +0000 UTC m=+1.180755521" Apr 28 00:15:46.835970 kubelet[2581]: E0428 00:15:46.835382 2581 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-7-n-51c70c830d\" already exists" pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" Apr 28 00:15:46.872440 kubelet[2581]: I0428 00:15:46.872179 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081-3-7-n-51c70c830d" podStartSLOduration=1.87216253 podStartE2EDuration="1.87216253s" podCreationTimestamp="2026-04-28 00:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-28 00:15:46.87091857 +0000 UTC m=+1.217012801" watchObservedRunningTime="2026-04-28 00:15:46.87216253 +0000 UTC m=+1.218256721" Apr 28 00:15:46.872440 kubelet[2581]: I0428 00:15:46.872373 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081-3-7-n-51c70c830d" podStartSLOduration=1.87236621 podStartE2EDuration="1.87236621s" podCreationTimestamp="2026-04-28 00:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-28 00:15:46.85058693 +0000 UTC m=+1.196681161" watchObservedRunningTime="2026-04-28 00:15:46.87236621 +0000 UTC m=+1.218460401" Apr 28 00:15:49.217515 kubelet[2581]: I0428 00:15:49.216997 2581 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 28 00:15:49.218738 containerd[1481]: time="2026-04-28T00:15:49.218630610Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 28 00:15:49.219770 kubelet[2581]: I0428 00:15:49.218928 2581 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 28 00:15:50.465300 systemd[1]: Created slice kubepods-besteffort-podd7d03447_6188_45a0_a6f8_748c943f49f9.slice - libcontainer container kubepods-besteffort-podd7d03447_6188_45a0_a6f8_748c943f49f9.slice. Apr 28 00:15:50.500694 kubelet[2581]: I0428 00:15:50.498924 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d7d03447-6188-45a0-a6f8-748c943f49f9-lib-modules\") pod \"kube-proxy-5g9bq\" (UID: \"d7d03447-6188-45a0-a6f8-748c943f49f9\") " pod="kube-system/kube-proxy-5g9bq" Apr 28 00:15:50.500694 kubelet[2581]: I0428 00:15:50.498971 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87dwt\" (UniqueName: \"kubernetes.io/projected/d7d03447-6188-45a0-a6f8-748c943f49f9-kube-api-access-87dwt\") pod \"kube-proxy-5g9bq\" (UID: \"d7d03447-6188-45a0-a6f8-748c943f49f9\") " pod="kube-system/kube-proxy-5g9bq" Apr 28 00:15:50.500694 kubelet[2581]: I0428 00:15:50.498990 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/d7d03447-6188-45a0-a6f8-748c943f49f9-kube-proxy\") pod \"kube-proxy-5g9bq\" (UID: \"d7d03447-6188-45a0-a6f8-748c943f49f9\") " pod="kube-system/kube-proxy-5g9bq" Apr 28 00:15:50.500694 kubelet[2581]: I0428 00:15:50.499005 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d7d03447-6188-45a0-a6f8-748c943f49f9-xtables-lock\") pod \"kube-proxy-5g9bq\" (UID: \"d7d03447-6188-45a0-a6f8-748c943f49f9\") " pod="kube-system/kube-proxy-5g9bq" Apr 28 00:15:50.612363 systemd[1]: Created slice kubepods-besteffort-podd21fdc26_3471_477d_a2b1_baa19a57994e.slice - libcontainer container kubepods-besteffort-podd21fdc26_3471_477d_a2b1_baa19a57994e.slice. Apr 28 00:15:50.699528 kubelet[2581]: I0428 00:15:50.699375 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njx8l\" (UniqueName: \"kubernetes.io/projected/d21fdc26-3471-477d-a2b1-baa19a57994e-kube-api-access-njx8l\") pod \"tigera-operator-8458958b4d-psz5j\" (UID: \"d21fdc26-3471-477d-a2b1-baa19a57994e\") " pod="tigera-operator/tigera-operator-8458958b4d-psz5j" Apr 28 00:15:50.699528 kubelet[2581]: I0428 00:15:50.699443 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/d21fdc26-3471-477d-a2b1-baa19a57994e-var-lib-calico\") pod \"tigera-operator-8458958b4d-psz5j\" (UID: \"d21fdc26-3471-477d-a2b1-baa19a57994e\") " pod="tigera-operator/tigera-operator-8458958b4d-psz5j" Apr 28 00:15:50.776589 containerd[1481]: time="2026-04-28T00:15:50.776435050Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5g9bq,Uid:d7d03447-6188-45a0-a6f8-748c943f49f9,Namespace:kube-system,Attempt:0,}" Apr 28 00:15:50.815706 containerd[1481]: time="2026-04-28T00:15:50.814882210Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:15:50.815706 containerd[1481]: time="2026-04-28T00:15:50.815006490Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:15:50.815706 containerd[1481]: time="2026-04-28T00:15:50.815033770Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:50.815706 containerd[1481]: time="2026-04-28T00:15:50.815227250Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:50.853482 systemd[1]: Started cri-containerd-825453ab30c433b67d83c0c8d9d87e1c915bbeee75a3205d7669f83d1d646ce0.scope - libcontainer container 825453ab30c433b67d83c0c8d9d87e1c915bbeee75a3205d7669f83d1d646ce0. Apr 28 00:15:50.877713 containerd[1481]: time="2026-04-28T00:15:50.877535290Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5g9bq,Uid:d7d03447-6188-45a0-a6f8-748c943f49f9,Namespace:kube-system,Attempt:0,} returns sandbox id \"825453ab30c433b67d83c0c8d9d87e1c915bbeee75a3205d7669f83d1d646ce0\"" Apr 28 00:15:50.884521 containerd[1481]: time="2026-04-28T00:15:50.884330010Z" level=info msg="CreateContainer within sandbox \"825453ab30c433b67d83c0c8d9d87e1c915bbeee75a3205d7669f83d1d646ce0\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 28 00:15:50.900053 containerd[1481]: time="2026-04-28T00:15:50.899880730Z" level=info msg="CreateContainer within sandbox \"825453ab30c433b67d83c0c8d9d87e1c915bbeee75a3205d7669f83d1d646ce0\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"f59b4a8e9f7ebd3b55d9885f6c3c4a2d5efcb5708b4cbc5c5fa20bfdefaf9649\"" Apr 28 00:15:50.901743 containerd[1481]: time="2026-04-28T00:15:50.901710890Z" level=info msg="StartContainer for \"f59b4a8e9f7ebd3b55d9885f6c3c4a2d5efcb5708b4cbc5c5fa20bfdefaf9649\"" Apr 28 00:15:50.918429 containerd[1481]: time="2026-04-28T00:15:50.918004810Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-8458958b4d-psz5j,Uid:d21fdc26-3471-477d-a2b1-baa19a57994e,Namespace:tigera-operator,Attempt:0,}" Apr 28 00:15:50.930902 systemd[1]: Started cri-containerd-f59b4a8e9f7ebd3b55d9885f6c3c4a2d5efcb5708b4cbc5c5fa20bfdefaf9649.scope - libcontainer container f59b4a8e9f7ebd3b55d9885f6c3c4a2d5efcb5708b4cbc5c5fa20bfdefaf9649. Apr 28 00:15:50.960999 containerd[1481]: time="2026-04-28T00:15:50.960847010Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:15:50.965406 containerd[1481]: time="2026-04-28T00:15:50.965321850Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:15:50.965647 containerd[1481]: time="2026-04-28T00:15:50.965577930Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:50.966290 containerd[1481]: time="2026-04-28T00:15:50.966055330Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:15:50.981767 containerd[1481]: time="2026-04-28T00:15:50.981681330Z" level=info msg="StartContainer for \"f59b4a8e9f7ebd3b55d9885f6c3c4a2d5efcb5708b4cbc5c5fa20bfdefaf9649\" returns successfully" Apr 28 00:15:50.996930 systemd[1]: Started cri-containerd-b54732ed22708aea8c9222cd6d89520b495adbbd3a7b041ad174a4e96accebb0.scope - libcontainer container b54732ed22708aea8c9222cd6d89520b495adbbd3a7b041ad174a4e96accebb0. Apr 28 00:15:51.037697 containerd[1481]: time="2026-04-28T00:15:51.037461410Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-8458958b4d-psz5j,Uid:d21fdc26-3471-477d-a2b1-baa19a57994e,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"b54732ed22708aea8c9222cd6d89520b495adbbd3a7b041ad174a4e96accebb0\"" Apr 28 00:15:51.041615 containerd[1481]: time="2026-04-28T00:15:51.041571530Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.8\"" Apr 28 00:15:51.634627 systemd[1]: run-containerd-runc-k8s.io-825453ab30c433b67d83c0c8d9d87e1c915bbeee75a3205d7669f83d1d646ce0-runc.e26riT.mount: Deactivated successfully. Apr 28 00:15:51.853142 kubelet[2581]: I0428 00:15:51.852350 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-5g9bq" podStartSLOduration=1.8523322100000001 podStartE2EDuration="1.85233221s" podCreationTimestamp="2026-04-28 00:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-28 00:15:51.85201985 +0000 UTC m=+6.198114041" watchObservedRunningTime="2026-04-28 00:15:51.85233221 +0000 UTC m=+6.198426401" Apr 28 00:15:52.657230 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount602915013.mount: Deactivated successfully. Apr 28 00:15:53.195589 containerd[1481]: time="2026-04-28T00:15:53.195517090Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:53.197829 containerd[1481]: time="2026-04-28T00:15:53.197302050Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.8: active requests=0, bytes read=24868969" Apr 28 00:15:53.197829 containerd[1481]: time="2026-04-28T00:15:53.197636930Z" level=info msg="ImageCreate event name:\"sha256:f37773829212e34063aa0c4c18558c40f2fc7ce0c68e8139b71af2ff71e26790\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:53.200527 containerd[1481]: time="2026-04-28T00:15:53.200477570Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:ce8eeaa3e60794610f3851ee06d296575f7c2efef1e3e1f8ac751a1d87ab979c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:15:53.202527 containerd[1481]: time="2026-04-28T00:15:53.201982930Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.8\" with image id \"sha256:f37773829212e34063aa0c4c18558c40f2fc7ce0c68e8139b71af2ff71e26790\", repo tag \"quay.io/tigera/operator:v1.40.8\", repo digest \"quay.io/tigera/operator@sha256:ce8eeaa3e60794610f3851ee06d296575f7c2efef1e3e1f8ac751a1d87ab979c\", size \"24864964\" in 2.16036072s" Apr 28 00:15:53.202527 containerd[1481]: time="2026-04-28T00:15:53.202034170Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.8\" returns image reference \"sha256:f37773829212e34063aa0c4c18558c40f2fc7ce0c68e8139b71af2ff71e26790\"" Apr 28 00:15:53.208984 containerd[1481]: time="2026-04-28T00:15:53.208922290Z" level=info msg="CreateContainer within sandbox \"b54732ed22708aea8c9222cd6d89520b495adbbd3a7b041ad174a4e96accebb0\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Apr 28 00:15:53.220376 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3242672647.mount: Deactivated successfully. Apr 28 00:15:53.226793 containerd[1481]: time="2026-04-28T00:15:53.226741730Z" level=info msg="CreateContainer within sandbox \"b54732ed22708aea8c9222cd6d89520b495adbbd3a7b041ad174a4e96accebb0\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0\"" Apr 28 00:15:53.230828 containerd[1481]: time="2026-04-28T00:15:53.229712050Z" level=info msg="StartContainer for \"e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0\"" Apr 28 00:15:53.265079 systemd[1]: Started cri-containerd-e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0.scope - libcontainer container e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0. Apr 28 00:15:53.299207 containerd[1481]: time="2026-04-28T00:15:53.299112010Z" level=info msg="StartContainer for \"e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0\" returns successfully" Apr 28 00:15:56.705025 kubelet[2581]: I0428 00:15:56.704947 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-8458958b4d-psz5j" podStartSLOduration=4.542756157 podStartE2EDuration="6.704927677s" podCreationTimestamp="2026-04-28 00:15:50 +0000 UTC" firstStartedPulling="2026-04-28 00:15:51.04096413 +0000 UTC m=+5.387058321" lastFinishedPulling="2026-04-28 00:15:53.20313565 +0000 UTC m=+7.549229841" observedRunningTime="2026-04-28 00:15:53.85726961 +0000 UTC m=+8.203363801" watchObservedRunningTime="2026-04-28 00:15:56.704927677 +0000 UTC m=+11.051021868" Apr 28 00:15:59.843780 sudo[1698]: pam_unix(sudo:session): session closed for user root Apr 28 00:15:59.862818 sshd[1695]: pam_unix(sshd:session): session closed for user core Apr 28 00:15:59.867851 systemd[1]: sshd@6-178.105.21.241:22-50.85.169.122:42942.service: Deactivated successfully. Apr 28 00:15:59.873784 systemd[1]: session-7.scope: Deactivated successfully. Apr 28 00:15:59.874419 systemd[1]: session-7.scope: Consumed 7.974s CPU time, 154.1M memory peak, 0B memory swap peak. Apr 28 00:15:59.877084 systemd-logind[1454]: Session 7 logged out. Waiting for processes to exit. Apr 28 00:15:59.879478 systemd-logind[1454]: Removed session 7. Apr 28 00:16:04.741071 systemd[1]: Created slice kubepods-besteffort-pod1eb30858_4ef6_41f6_aebd_3cd8140dc14d.slice - libcontainer container kubepods-besteffort-pod1eb30858_4ef6_41f6_aebd_3cd8140dc14d.slice. Apr 28 00:16:04.788841 kubelet[2581]: I0428 00:16:04.788237 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1eb30858-4ef6-41f6-aebd-3cd8140dc14d-tigera-ca-bundle\") pod \"calico-typha-774b7df59c-l9kwv\" (UID: \"1eb30858-4ef6-41f6-aebd-3cd8140dc14d\") " pod="calico-system/calico-typha-774b7df59c-l9kwv" Apr 28 00:16:04.788841 kubelet[2581]: I0428 00:16:04.788284 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/1eb30858-4ef6-41f6-aebd-3cd8140dc14d-typha-certs\") pod \"calico-typha-774b7df59c-l9kwv\" (UID: \"1eb30858-4ef6-41f6-aebd-3cd8140dc14d\") " pod="calico-system/calico-typha-774b7df59c-l9kwv" Apr 28 00:16:04.788841 kubelet[2581]: I0428 00:16:04.788303 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssbpv\" (UniqueName: \"kubernetes.io/projected/1eb30858-4ef6-41f6-aebd-3cd8140dc14d-kube-api-access-ssbpv\") pod \"calico-typha-774b7df59c-l9kwv\" (UID: \"1eb30858-4ef6-41f6-aebd-3cd8140dc14d\") " pod="calico-system/calico-typha-774b7df59c-l9kwv" Apr 28 00:16:04.878373 systemd[1]: Created slice kubepods-besteffort-pod3e613f8d_a73e_44ee_a75a_a037f873293f.slice - libcontainer container kubepods-besteffort-pod3e613f8d_a73e_44ee_a75a_a037f873293f.slice. Apr 28 00:16:04.890495 kubelet[2581]: I0428 00:16:04.889040 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-bpffs\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890495 kubelet[2581]: I0428 00:16:04.889090 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-cni-net-dir\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890495 kubelet[2581]: I0428 00:16:04.889107 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-lib-modules\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890495 kubelet[2581]: I0428 00:16:04.889123 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-xtables-lock\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890495 kubelet[2581]: I0428 00:16:04.889142 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/3e613f8d-a73e-44ee-a75a-a037f873293f-node-certs\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890750 kubelet[2581]: I0428 00:16:04.889216 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-cni-log-dir\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890750 kubelet[2581]: I0428 00:16:04.889234 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-cni-bin-dir\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890750 kubelet[2581]: I0428 00:16:04.889249 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-flexvol-driver-host\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890750 kubelet[2581]: I0428 00:16:04.889282 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-policysync\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890750 kubelet[2581]: I0428 00:16:04.889298 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-sys-fs\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890860 kubelet[2581]: I0428 00:16:04.889314 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e613f8d-a73e-44ee-a75a-a037f873293f-tigera-ca-bundle\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890860 kubelet[2581]: I0428 00:16:04.889331 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-var-lib-calico\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890860 kubelet[2581]: I0428 00:16:04.889345 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-var-run-calico\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890860 kubelet[2581]: I0428 00:16:04.889361 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntfv6\" (UniqueName: \"kubernetes.io/projected/3e613f8d-a73e-44ee-a75a-a037f873293f-kube-api-access-ntfv6\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.890860 kubelet[2581]: I0428 00:16:04.889386 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/3e613f8d-a73e-44ee-a75a-a037f873293f-nodeproc\") pod \"calico-node-x45hx\" (UID: \"3e613f8d-a73e-44ee-a75a-a037f873293f\") " pod="calico-system/calico-node-x45hx" Apr 28 00:16:04.984979 kubelet[2581]: E0428 00:16:04.984810 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:04.990902 kubelet[2581]: I0428 00:16:04.989523 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7335aad5-4b0b-4297-8eed-20f9df1b81e5-kubelet-dir\") pod \"csi-node-driver-dcppv\" (UID: \"7335aad5-4b0b-4297-8eed-20f9df1b81e5\") " pod="calico-system/csi-node-driver-dcppv" Apr 28 00:16:04.990902 kubelet[2581]: I0428 00:16:04.989568 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7335aad5-4b0b-4297-8eed-20f9df1b81e5-socket-dir\") pod \"csi-node-driver-dcppv\" (UID: \"7335aad5-4b0b-4297-8eed-20f9df1b81e5\") " pod="calico-system/csi-node-driver-dcppv" Apr 28 00:16:04.990902 kubelet[2581]: I0428 00:16:04.989617 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/7335aad5-4b0b-4297-8eed-20f9df1b81e5-varrun\") pod \"csi-node-driver-dcppv\" (UID: \"7335aad5-4b0b-4297-8eed-20f9df1b81e5\") " pod="calico-system/csi-node-driver-dcppv" Apr 28 00:16:04.990902 kubelet[2581]: I0428 00:16:04.989702 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7335aad5-4b0b-4297-8eed-20f9df1b81e5-registration-dir\") pod \"csi-node-driver-dcppv\" (UID: \"7335aad5-4b0b-4297-8eed-20f9df1b81e5\") " pod="calico-system/csi-node-driver-dcppv" Apr 28 00:16:04.990902 kubelet[2581]: I0428 00:16:04.989724 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx2gt\" (UniqueName: \"kubernetes.io/projected/7335aad5-4b0b-4297-8eed-20f9df1b81e5-kube-api-access-nx2gt\") pod \"csi-node-driver-dcppv\" (UID: \"7335aad5-4b0b-4297-8eed-20f9df1b81e5\") " pod="calico-system/csi-node-driver-dcppv" Apr 28 00:16:04.997203 kubelet[2581]: E0428 00:16:04.995789 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:04.997203 kubelet[2581]: W0428 00:16:04.995823 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:04.997203 kubelet[2581]: E0428 00:16:04.995850 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.000634 kubelet[2581]: E0428 00:16:05.000190 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.000634 kubelet[2581]: W0428 00:16:05.000214 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.000634 kubelet[2581]: E0428 00:16:05.000236 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.001741 kubelet[2581]: E0428 00:16:05.001311 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.001741 kubelet[2581]: W0428 00:16:05.001334 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.001741 kubelet[2581]: E0428 00:16:05.001352 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.002066 kubelet[2581]: E0428 00:16:05.001925 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.002066 kubelet[2581]: W0428 00:16:05.001940 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.002066 kubelet[2581]: E0428 00:16:05.001957 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.003843 kubelet[2581]: E0428 00:16:05.003005 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.003843 kubelet[2581]: W0428 00:16:05.003028 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.003843 kubelet[2581]: E0428 00:16:05.003043 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.004165 kubelet[2581]: E0428 00:16:05.003922 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.004165 kubelet[2581]: W0428 00:16:05.003996 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.004165 kubelet[2581]: E0428 00:16:05.004030 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.004889 kubelet[2581]: E0428 00:16:05.004856 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.004889 kubelet[2581]: W0428 00:16:05.004880 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.004973 kubelet[2581]: E0428 00:16:05.004896 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.005361 kubelet[2581]: E0428 00:16:05.005340 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.005361 kubelet[2581]: W0428 00:16:05.005357 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.005457 kubelet[2581]: E0428 00:16:05.005371 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.006075 kubelet[2581]: E0428 00:16:05.006035 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.006075 kubelet[2581]: W0428 00:16:05.006054 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.006075 kubelet[2581]: E0428 00:16:05.006066 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.006478 kubelet[2581]: E0428 00:16:05.006461 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.006478 kubelet[2581]: W0428 00:16:05.006478 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.006559 kubelet[2581]: E0428 00:16:05.006491 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.007976 kubelet[2581]: E0428 00:16:05.007947 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.007976 kubelet[2581]: W0428 00:16:05.007968 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.008082 kubelet[2581]: E0428 00:16:05.007983 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.008261 kubelet[2581]: E0428 00:16:05.008247 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.008261 kubelet[2581]: W0428 00:16:05.008261 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.008376 kubelet[2581]: E0428 00:16:05.008273 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.008482 kubelet[2581]: E0428 00:16:05.008471 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.008513 kubelet[2581]: W0428 00:16:05.008483 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.008513 kubelet[2581]: E0428 00:16:05.008492 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.009518 kubelet[2581]: E0428 00:16:05.009464 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.009518 kubelet[2581]: W0428 00:16:05.009513 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.009632 kubelet[2581]: E0428 00:16:05.009527 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.009841 kubelet[2581]: E0428 00:16:05.009823 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.009841 kubelet[2581]: W0428 00:16:05.009836 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.009841 kubelet[2581]: E0428 00:16:05.009847 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.010945 kubelet[2581]: E0428 00:16:05.010915 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.010945 kubelet[2581]: W0428 00:16:05.010935 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.010945 kubelet[2581]: E0428 00:16:05.010951 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.011449 kubelet[2581]: E0428 00:16:05.011423 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.011449 kubelet[2581]: W0428 00:16:05.011441 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.011449 kubelet[2581]: E0428 00:16:05.011454 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.013103 kubelet[2581]: E0428 00:16:05.012922 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.013103 kubelet[2581]: W0428 00:16:05.012943 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.013103 kubelet[2581]: E0428 00:16:05.012958 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.014249 kubelet[2581]: E0428 00:16:05.014055 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.014249 kubelet[2581]: W0428 00:16:05.014075 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.014249 kubelet[2581]: E0428 00:16:05.014090 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.015182 kubelet[2581]: E0428 00:16:05.015113 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.015182 kubelet[2581]: W0428 00:16:05.015132 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.015182 kubelet[2581]: E0428 00:16:05.015183 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.015945 kubelet[2581]: E0428 00:16:05.015922 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.016762 kubelet[2581]: W0428 00:16:05.016725 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.016762 kubelet[2581]: E0428 00:16:05.016760 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.017381 kubelet[2581]: E0428 00:16:05.017344 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.017381 kubelet[2581]: W0428 00:16:05.017373 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.017381 kubelet[2581]: E0428 00:16:05.017387 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.017936 kubelet[2581]: E0428 00:16:05.017917 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.017936 kubelet[2581]: W0428 00:16:05.017932 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.018043 kubelet[2581]: E0428 00:16:05.017943 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.018225 kubelet[2581]: E0428 00:16:05.018210 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.018225 kubelet[2581]: W0428 00:16:05.018225 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.018278 kubelet[2581]: E0428 00:16:05.018237 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.018755 kubelet[2581]: E0428 00:16:05.018732 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.018755 kubelet[2581]: W0428 00:16:05.018752 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.018853 kubelet[2581]: E0428 00:16:05.018765 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.019650 kubelet[2581]: E0428 00:16:05.019628 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.019650 kubelet[2581]: W0428 00:16:05.019650 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.019882 kubelet[2581]: E0428 00:16:05.019679 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.039856 kubelet[2581]: E0428 00:16:05.039805 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.039856 kubelet[2581]: W0428 00:16:05.039836 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.039856 kubelet[2581]: E0428 00:16:05.039860 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.047306 containerd[1481]: time="2026-04-28T00:16:05.047248779Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-774b7df59c-l9kwv,Uid:1eb30858-4ef6-41f6-aebd-3cd8140dc14d,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:05.080733 containerd[1481]: time="2026-04-28T00:16:05.080571713Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:05.080733 containerd[1481]: time="2026-04-28T00:16:05.080650994Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:05.081657 containerd[1481]: time="2026-04-28T00:16:05.080665634Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:05.081657 containerd[1481]: time="2026-04-28T00:16:05.081605887Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:05.091131 kubelet[2581]: E0428 00:16:05.091094 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.091131 kubelet[2581]: W0428 00:16:05.091118 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.091131 kubelet[2581]: E0428 00:16:05.091141 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.092345 kubelet[2581]: E0428 00:16:05.091632 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.092345 kubelet[2581]: W0428 00:16:05.091646 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.092345 kubelet[2581]: E0428 00:16:05.091679 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.093857 kubelet[2581]: E0428 00:16:05.093827 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.093857 kubelet[2581]: W0428 00:16:05.093847 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.093857 kubelet[2581]: E0428 00:16:05.093863 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.094958 kubelet[2581]: E0428 00:16:05.094851 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.094958 kubelet[2581]: W0428 00:16:05.094879 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.094958 kubelet[2581]: E0428 00:16:05.094894 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.095431 kubelet[2581]: E0428 00:16:05.095415 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.095881 kubelet[2581]: W0428 00:16:05.095749 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.095881 kubelet[2581]: E0428 00:16:05.095778 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.096867 kubelet[2581]: E0428 00:16:05.096704 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.096867 kubelet[2581]: W0428 00:16:05.096720 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.096867 kubelet[2581]: E0428 00:16:05.096732 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.098821 kubelet[2581]: E0428 00:16:05.098336 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.098821 kubelet[2581]: W0428 00:16:05.098533 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.098821 kubelet[2581]: E0428 00:16:05.098559 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.100991 kubelet[2581]: E0428 00:16:05.100602 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.100991 kubelet[2581]: W0428 00:16:05.100621 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.100991 kubelet[2581]: E0428 00:16:05.100640 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.102004 kubelet[2581]: E0428 00:16:05.101949 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.102322 kubelet[2581]: W0428 00:16:05.102135 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.102689 kubelet[2581]: E0428 00:16:05.102496 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.104613 kubelet[2581]: E0428 00:16:05.104132 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.104613 kubelet[2581]: W0428 00:16:05.104197 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.104613 kubelet[2581]: E0428 00:16:05.104218 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.105511 kubelet[2581]: E0428 00:16:05.105222 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.105511 kubelet[2581]: W0428 00:16:05.105240 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.105511 kubelet[2581]: E0428 00:16:05.105257 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.106136 kubelet[2581]: E0428 00:16:05.106114 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.106620 kubelet[2581]: W0428 00:16:05.106302 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.106620 kubelet[2581]: E0428 00:16:05.106327 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.106900 kubelet[2581]: E0428 00:16:05.106885 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.108922 kubelet[2581]: W0428 00:16:05.108724 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.108922 kubelet[2581]: E0428 00:16:05.108760 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.109088 kubelet[2581]: E0428 00:16:05.109076 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.109163 kubelet[2581]: W0428 00:16:05.109141 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.109213 kubelet[2581]: E0428 00:16:05.109203 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.109510 kubelet[2581]: E0428 00:16:05.109494 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.109603 kubelet[2581]: W0428 00:16:05.109590 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.109776 kubelet[2581]: E0428 00:16:05.109650 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.109888 kubelet[2581]: E0428 00:16:05.109875 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.109958 kubelet[2581]: W0428 00:16:05.109946 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.110016 kubelet[2581]: E0428 00:16:05.110005 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.110272 kubelet[2581]: E0428 00:16:05.110255 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.110446 kubelet[2581]: W0428 00:16:05.110344 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.110446 kubelet[2581]: E0428 00:16:05.110361 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.110575 kubelet[2581]: E0428 00:16:05.110563 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.110628 kubelet[2581]: W0428 00:16:05.110618 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.110706 kubelet[2581]: E0428 00:16:05.110691 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.111277 kubelet[2581]: E0428 00:16:05.111197 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.111589 kubelet[2581]: W0428 00:16:05.111435 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.111589 kubelet[2581]: E0428 00:16:05.111458 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.111941 systemd[1]: Started cri-containerd-185ad22cee88b21e1dbcbcfa3281af78aa92602fdb9380a4d1ddda718adc0f54.scope - libcontainer container 185ad22cee88b21e1dbcbcfa3281af78aa92602fdb9380a4d1ddda718adc0f54. Apr 28 00:16:05.113405 kubelet[2581]: E0428 00:16:05.113361 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.113405 kubelet[2581]: W0428 00:16:05.113402 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.113482 kubelet[2581]: E0428 00:16:05.113417 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.114568 kubelet[2581]: E0428 00:16:05.114364 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.114568 kubelet[2581]: W0428 00:16:05.114392 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.114568 kubelet[2581]: E0428 00:16:05.114404 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.115068 kubelet[2581]: E0428 00:16:05.114958 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.115068 kubelet[2581]: W0428 00:16:05.114977 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.115068 kubelet[2581]: E0428 00:16:05.114993 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.116407 kubelet[2581]: E0428 00:16:05.116215 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.116407 kubelet[2581]: W0428 00:16:05.116241 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.116407 kubelet[2581]: E0428 00:16:05.116260 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.117198 kubelet[2581]: E0428 00:16:05.117174 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.117198 kubelet[2581]: W0428 00:16:05.117194 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.117198 kubelet[2581]: E0428 00:16:05.117216 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.118029 kubelet[2581]: E0428 00:16:05.117966 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.118029 kubelet[2581]: W0428 00:16:05.117985 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.118029 kubelet[2581]: E0428 00:16:05.117999 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.129568 kubelet[2581]: E0428 00:16:05.129477 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:05.129568 kubelet[2581]: W0428 00:16:05.129500 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:05.129568 kubelet[2581]: E0428 00:16:05.129520 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:05.160544 containerd[1481]: time="2026-04-28T00:16:05.160482602Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-774b7df59c-l9kwv,Uid:1eb30858-4ef6-41f6-aebd-3cd8140dc14d,Namespace:calico-system,Attempt:0,} returns sandbox id \"185ad22cee88b21e1dbcbcfa3281af78aa92602fdb9380a4d1ddda718adc0f54\"" Apr 28 00:16:05.163752 containerd[1481]: time="2026-04-28T00:16:05.163036477Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.5\"" Apr 28 00:16:05.184575 containerd[1481]: time="2026-04-28T00:16:05.184522690Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-x45hx,Uid:3e613f8d-a73e-44ee-a75a-a037f873293f,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:05.221330 containerd[1481]: time="2026-04-28T00:16:05.221042868Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:05.221330 containerd[1481]: time="2026-04-28T00:16:05.221117829Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:05.221330 containerd[1481]: time="2026-04-28T00:16:05.221134349Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:05.221330 containerd[1481]: time="2026-04-28T00:16:05.221244071Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:05.243906 systemd[1]: Started cri-containerd-d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb.scope - libcontainer container d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb. Apr 28 00:16:05.273829 containerd[1481]: time="2026-04-28T00:16:05.273472823Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-x45hx,Uid:3e613f8d-a73e-44ee-a75a-a037f873293f,Namespace:calico-system,Attempt:0,} returns sandbox id \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\"" Apr 28 00:16:06.707780 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2014721799.mount: Deactivated successfully. Apr 28 00:16:06.777881 kubelet[2581]: E0428 00:16:06.777826 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:07.206816 containerd[1481]: time="2026-04-28T00:16:07.205765900Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:07.208183 containerd[1481]: time="2026-04-28T00:16:07.208130368Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.5: active requests=0, bytes read=32841445" Apr 28 00:16:07.209761 containerd[1481]: time="2026-04-28T00:16:07.209511265Z" level=info msg="ImageCreate event name:\"sha256:265c145eea96693e7abfe97a68dee913c8e656947f5708c28e4e866d3809b4c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:07.211938 containerd[1481]: time="2026-04-28T00:16:07.211855213Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:76afd8f80569b3bf783991ce5348294319cefa6d6cca127710d0e068096048a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:07.213351 containerd[1481]: time="2026-04-28T00:16:07.213179309Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.5\" with image id \"sha256:265c145eea96693e7abfe97a68dee913c8e656947f5708c28e4e866d3809b4c9\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:76afd8f80569b3bf783991ce5348294319cefa6d6cca127710d0e068096048a6\", size \"32841299\" in 2.050094631s" Apr 28 00:16:07.213351 containerd[1481]: time="2026-04-28T00:16:07.213230309Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.5\" returns image reference \"sha256:265c145eea96693e7abfe97a68dee913c8e656947f5708c28e4e866d3809b4c9\"" Apr 28 00:16:07.215359 containerd[1481]: time="2026-04-28T00:16:07.215315814Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\"" Apr 28 00:16:07.232997 containerd[1481]: time="2026-04-28T00:16:07.232818024Z" level=info msg="CreateContainer within sandbox \"185ad22cee88b21e1dbcbcfa3281af78aa92602fdb9380a4d1ddda718adc0f54\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Apr 28 00:16:07.253037 containerd[1481]: time="2026-04-28T00:16:07.252891065Z" level=info msg="CreateContainer within sandbox \"185ad22cee88b21e1dbcbcfa3281af78aa92602fdb9380a4d1ddda718adc0f54\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"d6959674a41f1c5a553db03e57ba095f0b488f2df4b9bcec58ba5f1cb7808270\"" Apr 28 00:16:07.254937 containerd[1481]: time="2026-04-28T00:16:07.254908649Z" level=info msg="StartContainer for \"d6959674a41f1c5a553db03e57ba095f0b488f2df4b9bcec58ba5f1cb7808270\"" Apr 28 00:16:07.299297 systemd[1]: Started cri-containerd-d6959674a41f1c5a553db03e57ba095f0b488f2df4b9bcec58ba5f1cb7808270.scope - libcontainer container d6959674a41f1c5a553db03e57ba095f0b488f2df4b9bcec58ba5f1cb7808270. Apr 28 00:16:07.373500 containerd[1481]: time="2026-04-28T00:16:07.373452230Z" level=info msg="StartContainer for \"d6959674a41f1c5a553db03e57ba095f0b488f2df4b9bcec58ba5f1cb7808270\" returns successfully" Apr 28 00:16:07.897206 kubelet[2581]: I0428 00:16:07.897032 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-774b7df59c-l9kwv" podStartSLOduration=1.844885728 podStartE2EDuration="3.897016144s" podCreationTimestamp="2026-04-28 00:16:04 +0000 UTC" firstStartedPulling="2026-04-28 00:16:05.162427069 +0000 UTC m=+19.508521260" lastFinishedPulling="2026-04-28 00:16:07.214557525 +0000 UTC m=+21.560651676" observedRunningTime="2026-04-28 00:16:07.895486725 +0000 UTC m=+22.241580916" watchObservedRunningTime="2026-04-28 00:16:07.897016144 +0000 UTC m=+22.243110335" Apr 28 00:16:07.909637 kubelet[2581]: E0428 00:16:07.909397 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.909637 kubelet[2581]: W0428 00:16:07.909435 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.909637 kubelet[2581]: E0428 00:16:07.909466 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.910380 kubelet[2581]: E0428 00:16:07.910041 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.910380 kubelet[2581]: W0428 00:16:07.910067 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.910380 kubelet[2581]: E0428 00:16:07.910134 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.910775 kubelet[2581]: E0428 00:16:07.910748 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.910919 kubelet[2581]: W0428 00:16:07.910895 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.911265 kubelet[2581]: E0428 00:16:07.911016 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.911535 kubelet[2581]: E0428 00:16:07.911452 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.911689 kubelet[2581]: W0428 00:16:07.911644 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.911848 kubelet[2581]: E0428 00:16:07.911822 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.912550 kubelet[2581]: E0428 00:16:07.912285 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.912550 kubelet[2581]: W0428 00:16:07.912315 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.912550 kubelet[2581]: E0428 00:16:07.912341 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.912987 kubelet[2581]: E0428 00:16:07.912958 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.913124 kubelet[2581]: W0428 00:16:07.913100 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.913551 kubelet[2581]: E0428 00:16:07.913276 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.913835 kubelet[2581]: E0428 00:16:07.913808 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.913987 kubelet[2581]: W0428 00:16:07.913960 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.914320 kubelet[2581]: E0428 00:16:07.914117 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.914560 kubelet[2581]: E0428 00:16:07.914534 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.914748 kubelet[2581]: W0428 00:16:07.914720 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.914904 kubelet[2581]: E0428 00:16:07.914877 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.915705 kubelet[2581]: E0428 00:16:07.915453 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.915705 kubelet[2581]: W0428 00:16:07.915481 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.915705 kubelet[2581]: E0428 00:16:07.915505 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.916028 kubelet[2581]: E0428 00:16:07.916012 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.916208 kubelet[2581]: W0428 00:16:07.916083 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.916208 kubelet[2581]: E0428 00:16:07.916099 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.916349 kubelet[2581]: E0428 00:16:07.916337 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.916406 kubelet[2581]: W0428 00:16:07.916395 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.916461 kubelet[2581]: E0428 00:16:07.916449 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.916783 kubelet[2581]: E0428 00:16:07.916769 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.916974 kubelet[2581]: W0428 00:16:07.916855 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.916974 kubelet[2581]: E0428 00:16:07.916871 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.917185 kubelet[2581]: E0428 00:16:07.917171 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.917297 kubelet[2581]: W0428 00:16:07.917284 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.917362 kubelet[2581]: E0428 00:16:07.917349 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.917769 kubelet[2581]: E0428 00:16:07.917630 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.917769 kubelet[2581]: W0428 00:16:07.917657 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.917769 kubelet[2581]: E0428 00:16:07.917684 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.917972 kubelet[2581]: E0428 00:16:07.917959 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.918032 kubelet[2581]: W0428 00:16:07.918022 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.918136 kubelet[2581]: E0428 00:16:07.918076 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.934335 kubelet[2581]: E0428 00:16:07.933525 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.934335 kubelet[2581]: W0428 00:16:07.933551 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.934335 kubelet[2581]: E0428 00:16:07.933576 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.934335 kubelet[2581]: E0428 00:16:07.933981 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.934335 kubelet[2581]: W0428 00:16:07.933998 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.934335 kubelet[2581]: E0428 00:16:07.934014 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.935656 kubelet[2581]: E0428 00:16:07.935019 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.935656 kubelet[2581]: W0428 00:16:07.935037 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.935656 kubelet[2581]: E0428 00:16:07.935055 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.937871 kubelet[2581]: E0428 00:16:07.937589 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.937871 kubelet[2581]: W0428 00:16:07.937619 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.937871 kubelet[2581]: E0428 00:16:07.937642 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.938770 kubelet[2581]: E0428 00:16:07.938248 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.938770 kubelet[2581]: W0428 00:16:07.938273 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.938770 kubelet[2581]: E0428 00:16:07.938290 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.938770 kubelet[2581]: E0428 00:16:07.938495 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.938770 kubelet[2581]: W0428 00:16:07.938502 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.938770 kubelet[2581]: E0428 00:16:07.938513 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.939254 kubelet[2581]: E0428 00:16:07.939107 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.939254 kubelet[2581]: W0428 00:16:07.939136 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.939254 kubelet[2581]: E0428 00:16:07.939164 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.940945 kubelet[2581]: E0428 00:16:07.939625 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.940945 kubelet[2581]: W0428 00:16:07.939638 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.940945 kubelet[2581]: E0428 00:16:07.939651 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.940945 kubelet[2581]: E0428 00:16:07.939950 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.940945 kubelet[2581]: W0428 00:16:07.939959 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.940945 kubelet[2581]: E0428 00:16:07.939968 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.940945 kubelet[2581]: E0428 00:16:07.940173 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.940945 kubelet[2581]: W0428 00:16:07.940182 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.940945 kubelet[2581]: E0428 00:16:07.940192 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.940945 kubelet[2581]: E0428 00:16:07.940361 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.941265 kubelet[2581]: W0428 00:16:07.940371 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.941265 kubelet[2581]: E0428 00:16:07.940380 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.941265 kubelet[2581]: E0428 00:16:07.940581 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.941265 kubelet[2581]: W0428 00:16:07.940589 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.941265 kubelet[2581]: E0428 00:16:07.940598 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.945347 kubelet[2581]: E0428 00:16:07.944347 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.945347 kubelet[2581]: W0428 00:16:07.944381 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.945347 kubelet[2581]: E0428 00:16:07.944403 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.945347 kubelet[2581]: E0428 00:16:07.944729 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.945347 kubelet[2581]: W0428 00:16:07.944739 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.945347 kubelet[2581]: E0428 00:16:07.944748 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.945347 kubelet[2581]: E0428 00:16:07.945093 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.945347 kubelet[2581]: W0428 00:16:07.945100 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.945347 kubelet[2581]: E0428 00:16:07.945110 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.946203 kubelet[2581]: E0428 00:16:07.945739 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.946203 kubelet[2581]: W0428 00:16:07.945755 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.946203 kubelet[2581]: E0428 00:16:07.945767 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.946203 kubelet[2581]: E0428 00:16:07.946134 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.946203 kubelet[2581]: W0428 00:16:07.946144 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.946203 kubelet[2581]: E0428 00:16:07.946167 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:07.947139 kubelet[2581]: E0428 00:16:07.947009 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:07.947139 kubelet[2581]: W0428 00:16:07.947049 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:07.947139 kubelet[2581]: E0428 00:16:07.947077 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.779346 kubelet[2581]: E0428 00:16:08.778345 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:08.881751 kubelet[2581]: I0428 00:16:08.881720 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 28 00:16:08.917378 containerd[1481]: time="2026-04-28T00:16:08.917294285Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:08.919052 containerd[1481]: time="2026-04-28T00:16:08.918984944Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5: active requests=0, bytes read=4404646" Apr 28 00:16:08.920694 containerd[1481]: time="2026-04-28T00:16:08.920418960Z" level=info msg="ImageCreate event name:\"sha256:3867b4c2eaa3321472d76c87dc2b4f8d6cdd45473f2138098e7ef206bc16d421\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:08.924845 containerd[1481]: time="2026-04-28T00:16:08.924017400Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:df00fee6895ac073066d91243f29733e71f479317cacef49d50c244bb2d21ea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:08.924845 containerd[1481]: time="2026-04-28T00:16:08.924735408Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\" with image id \"sha256:3867b4c2eaa3321472d76c87dc2b4f8d6cdd45473f2138098e7ef206bc16d421\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:df00fee6895ac073066d91243f29733e71f479317cacef49d50c244bb2d21ea1\", size \"6980245\" in 1.709376473s" Apr 28 00:16:08.924845 containerd[1481]: time="2026-04-28T00:16:08.924766688Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.5\" returns image reference \"sha256:3867b4c2eaa3321472d76c87dc2b4f8d6cdd45473f2138098e7ef206bc16d421\"" Apr 28 00:16:08.925500 kubelet[2581]: E0428 00:16:08.925391 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.925500 kubelet[2581]: W0428 00:16:08.925409 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.925500 kubelet[2581]: E0428 00:16:08.925428 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.925952 kubelet[2581]: E0428 00:16:08.925938 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.926018 kubelet[2581]: W0428 00:16:08.926005 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.926762 kubelet[2581]: E0428 00:16:08.926536 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.926898 kubelet[2581]: E0428 00:16:08.926883 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.927066 kubelet[2581]: W0428 00:16:08.926949 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.927066 kubelet[2581]: E0428 00:16:08.926965 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.928243 kubelet[2581]: E0428 00:16:08.928090 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.928243 kubelet[2581]: W0428 00:16:08.928120 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.928243 kubelet[2581]: E0428 00:16:08.928135 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.929423 kubelet[2581]: E0428 00:16:08.929392 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.930310 kubelet[2581]: W0428 00:16:08.930041 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.930310 kubelet[2581]: E0428 00:16:08.930064 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.930499 kubelet[2581]: E0428 00:16:08.930483 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.930657 kubelet[2581]: W0428 00:16:08.930551 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.930657 kubelet[2581]: E0428 00:16:08.930569 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.930990 kubelet[2581]: E0428 00:16:08.930904 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.930990 kubelet[2581]: W0428 00:16:08.930918 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.930990 kubelet[2581]: E0428 00:16:08.930931 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.931699 kubelet[2581]: E0428 00:16:08.931622 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.931699 kubelet[2581]: W0428 00:16:08.931638 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.931699 kubelet[2581]: E0428 00:16:08.931651 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.932174 kubelet[2581]: E0428 00:16:08.932137 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.932174 kubelet[2581]: W0428 00:16:08.932169 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.932238 kubelet[2581]: E0428 00:16:08.932183 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.932422 kubelet[2581]: E0428 00:16:08.932408 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.932422 kubelet[2581]: W0428 00:16:08.932423 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.932479 kubelet[2581]: E0428 00:16:08.932433 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.932646 kubelet[2581]: E0428 00:16:08.932630 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.932646 kubelet[2581]: W0428 00:16:08.932646 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.933022 kubelet[2581]: E0428 00:16:08.932656 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.933531 kubelet[2581]: E0428 00:16:08.933509 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.933531 kubelet[2581]: W0428 00:16:08.933530 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.933628 kubelet[2581]: E0428 00:16:08.933545 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.933918 kubelet[2581]: E0428 00:16:08.933787 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.933918 kubelet[2581]: W0428 00:16:08.933801 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.933918 kubelet[2581]: E0428 00:16:08.933811 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.934150 kubelet[2581]: E0428 00:16:08.934057 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.934150 kubelet[2581]: W0428 00:16:08.934067 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.934150 kubelet[2581]: E0428 00:16:08.934078 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.934260 kubelet[2581]: E0428 00:16:08.934249 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.934260 kubelet[2581]: W0428 00:16:08.934258 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.934311 kubelet[2581]: E0428 00:16:08.934267 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.935131 containerd[1481]: time="2026-04-28T00:16:08.934851562Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Apr 28 00:16:08.941611 kubelet[2581]: E0428 00:16:08.941460 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.941611 kubelet[2581]: W0428 00:16:08.941485 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.941611 kubelet[2581]: E0428 00:16:08.941504 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.942048 kubelet[2581]: E0428 00:16:08.941940 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.942048 kubelet[2581]: W0428 00:16:08.941973 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.942048 kubelet[2581]: E0428 00:16:08.941986 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.942339 kubelet[2581]: E0428 00:16:08.942321 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.942404 kubelet[2581]: W0428 00:16:08.942340 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.942404 kubelet[2581]: E0428 00:16:08.942354 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.942635 kubelet[2581]: E0428 00:16:08.942621 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.942635 kubelet[2581]: W0428 00:16:08.942635 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.942738 kubelet[2581]: E0428 00:16:08.942647 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.942895 kubelet[2581]: E0428 00:16:08.942880 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.942943 kubelet[2581]: W0428 00:16:08.942894 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.942943 kubelet[2581]: E0428 00:16:08.942905 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.943810 kubelet[2581]: E0428 00:16:08.943785 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.943810 kubelet[2581]: W0428 00:16:08.943804 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.944188 kubelet[2581]: E0428 00:16:08.943818 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.945833 kubelet[2581]: E0428 00:16:08.945724 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.945833 kubelet[2581]: W0428 00:16:08.945743 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.945833 kubelet[2581]: E0428 00:16:08.945768 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.946199 kubelet[2581]: E0428 00:16:08.946176 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.946199 kubelet[2581]: W0428 00:16:08.946195 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.946289 kubelet[2581]: E0428 00:16:08.946210 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.946462 kubelet[2581]: E0428 00:16:08.946443 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.946462 kubelet[2581]: W0428 00:16:08.946460 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.946569 kubelet[2581]: E0428 00:16:08.946471 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.946647 kubelet[2581]: E0428 00:16:08.946631 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.946647 kubelet[2581]: W0428 00:16:08.946644 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.946737 kubelet[2581]: E0428 00:16:08.946654 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.949377 kubelet[2581]: E0428 00:16:08.949342 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.949377 kubelet[2581]: W0428 00:16:08.949371 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.949516 kubelet[2581]: E0428 00:16:08.949395 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.951557 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3096494816.mount: Deactivated successfully. Apr 28 00:16:08.952082 kubelet[2581]: E0428 00:16:08.951595 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.952082 kubelet[2581]: W0428 00:16:08.951613 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.952082 kubelet[2581]: E0428 00:16:08.951630 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.953114 kubelet[2581]: E0428 00:16:08.952800 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.953114 kubelet[2581]: W0428 00:16:08.952829 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.953114 kubelet[2581]: E0428 00:16:08.953023 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.953642 kubelet[2581]: E0428 00:16:08.953603 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.953642 kubelet[2581]: W0428 00:16:08.953617 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.953642 kubelet[2581]: E0428 00:16:08.953630 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.954861 kubelet[2581]: E0428 00:16:08.954445 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.954861 kubelet[2581]: W0428 00:16:08.954460 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.954861 kubelet[2581]: E0428 00:16:08.954704 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.955068 kubelet[2581]: E0428 00:16:08.955054 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.955474 kubelet[2581]: W0428 00:16:08.955133 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.955474 kubelet[2581]: E0428 00:16:08.955170 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.955824 kubelet[2581]: E0428 00:16:08.955807 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.955912 kubelet[2581]: W0428 00:16:08.955899 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.955967 kubelet[2581]: E0428 00:16:08.955956 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.956439 containerd[1481]: time="2026-04-28T00:16:08.956205362Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f\"" Apr 28 00:16:08.958086 kubelet[2581]: E0428 00:16:08.957546 2581 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 28 00:16:08.958959 kubelet[2581]: W0428 00:16:08.958220 2581 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 28 00:16:08.958959 kubelet[2581]: E0428 00:16:08.958246 2581 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 28 00:16:08.959191 containerd[1481]: time="2026-04-28T00:16:08.958430227Z" level=info msg="StartContainer for \"ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f\"" Apr 28 00:16:08.997069 systemd[1]: Started cri-containerd-ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f.scope - libcontainer container ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f. Apr 28 00:16:09.029646 containerd[1481]: time="2026-04-28T00:16:09.029464485Z" level=info msg="StartContainer for \"ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f\" returns successfully" Apr 28 00:16:09.047810 systemd[1]: cri-containerd-ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f.scope: Deactivated successfully. Apr 28 00:16:09.071187 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f-rootfs.mount: Deactivated successfully. Apr 28 00:16:09.166335 containerd[1481]: time="2026-04-28T00:16:09.166198565Z" level=info msg="shim disconnected" id=ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f namespace=k8s.io Apr 28 00:16:09.166335 containerd[1481]: time="2026-04-28T00:16:09.166263565Z" level=warning msg="cleaning up after shim disconnected" id=ade0c0e7a1f0ad64daffba5dcbb48367bb9da52e09642eb95f79903c6f55921f namespace=k8s.io Apr 28 00:16:09.166335 containerd[1481]: time="2026-04-28T00:16:09.166274406Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 28 00:16:09.895474 containerd[1481]: time="2026-04-28T00:16:09.895138162Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.5\"" Apr 28 00:16:10.779193 kubelet[2581]: E0428 00:16:10.779067 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:12.779729 kubelet[2581]: E0428 00:16:12.778653 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:14.223032 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3905874501.mount: Deactivated successfully. Apr 28 00:16:14.254235 containerd[1481]: time="2026-04-28T00:16:14.254139233Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:14.256562 containerd[1481]: time="2026-04-28T00:16:14.256514451Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.5: active requests=0, bytes read=153029581" Apr 28 00:16:14.257817 containerd[1481]: time="2026-04-28T00:16:14.257765581Z" level=info msg="ImageCreate event name:\"sha256:5a8f90ba0ad45873b37c9c512d6391f35086ced5c27f20cfc5c45f777f9941b3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:14.261805 containerd[1481]: time="2026-04-28T00:16:14.260953485Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e2426b97a645ed620e0f4035d594f2f3344b0547cd3dc3458f45e06d5cebdad7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:14.261805 containerd[1481]: time="2026-04-28T00:16:14.261423529Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.5\" with image id \"sha256:5a8f90ba0ad45873b37c9c512d6391f35086ced5c27f20cfc5c45f777f9941b3\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e2426b97a645ed620e0f4035d594f2f3344b0547cd3dc3458f45e06d5cebdad7\", size \"153029443\" in 4.366213806s" Apr 28 00:16:14.261805 containerd[1481]: time="2026-04-28T00:16:14.261456289Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.5\" returns image reference \"sha256:5a8f90ba0ad45873b37c9c512d6391f35086ced5c27f20cfc5c45f777f9941b3\"" Apr 28 00:16:14.267955 containerd[1481]: time="2026-04-28T00:16:14.267866418Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Apr 28 00:16:14.287098 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount391275315.mount: Deactivated successfully. Apr 28 00:16:14.290847 containerd[1481]: time="2026-04-28T00:16:14.290800993Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602\"" Apr 28 00:16:14.293120 containerd[1481]: time="2026-04-28T00:16:14.291475398Z" level=info msg="StartContainer for \"3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602\"" Apr 28 00:16:14.336929 systemd[1]: Started cri-containerd-3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602.scope - libcontainer container 3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602. Apr 28 00:16:14.370926 containerd[1481]: time="2026-04-28T00:16:14.370877364Z" level=info msg="StartContainer for \"3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602\" returns successfully" Apr 28 00:16:14.482719 systemd[1]: cri-containerd-3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602.scope: Deactivated successfully. Apr 28 00:16:14.646407 containerd[1481]: time="2026-04-28T00:16:14.646181583Z" level=info msg="shim disconnected" id=3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602 namespace=k8s.io Apr 28 00:16:14.646407 containerd[1481]: time="2026-04-28T00:16:14.646245544Z" level=warning msg="cleaning up after shim disconnected" id=3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602 namespace=k8s.io Apr 28 00:16:14.646407 containerd[1481]: time="2026-04-28T00:16:14.646254144Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 28 00:16:14.778734 kubelet[2581]: E0428 00:16:14.778555 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:14.906346 containerd[1481]: time="2026-04-28T00:16:14.906187767Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.5\"" Apr 28 00:16:15.224902 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3a034311939a2f9885fc2bf1179504be6254c3e6ebc2cd3c3e2d861f6a2c4602-rootfs.mount: Deactivated successfully. Apr 28 00:16:16.779725 kubelet[2581]: E0428 00:16:16.778879 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:17.534340 containerd[1481]: time="2026-04-28T00:16:17.534285096Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:17.536944 containerd[1481]: time="2026-04-28T00:16:17.536912872Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.5: active requests=0, bytes read=62266008" Apr 28 00:16:17.538610 containerd[1481]: time="2026-04-28T00:16:17.538578923Z" level=info msg="ImageCreate event name:\"sha256:0636f5f0fe5e716fd01c674abaaef326193e41f0291d3a9b0ce572a82500c211\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:17.541766 containerd[1481]: time="2026-04-28T00:16:17.541728503Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:ea8a6b721af629c1dab2e1559b93cd843d9a4b640726115380fc23cf47e83232\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:17.543061 containerd[1481]: time="2026-04-28T00:16:17.543028591Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.5\" with image id \"sha256:0636f5f0fe5e716fd01c674abaaef326193e41f0291d3a9b0ce572a82500c211\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:ea8a6b721af629c1dab2e1559b93cd843d9a4b640726115380fc23cf47e83232\", size \"64841647\" in 2.636737824s" Apr 28 00:16:17.543177 containerd[1481]: time="2026-04-28T00:16:17.543160272Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.5\" returns image reference \"sha256:0636f5f0fe5e716fd01c674abaaef326193e41f0291d3a9b0ce572a82500c211\"" Apr 28 00:16:17.548539 containerd[1481]: time="2026-04-28T00:16:17.548499385Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Apr 28 00:16:17.567605 containerd[1481]: time="2026-04-28T00:16:17.567528185Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78\"" Apr 28 00:16:17.568713 containerd[1481]: time="2026-04-28T00:16:17.568471791Z" level=info msg="StartContainer for \"dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78\"" Apr 28 00:16:17.606872 systemd[1]: Started cri-containerd-dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78.scope - libcontainer container dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78. Apr 28 00:16:17.642856 containerd[1481]: time="2026-04-28T00:16:17.642806538Z" level=info msg="StartContainer for \"dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78\" returns successfully" Apr 28 00:16:18.217885 systemd[1]: cri-containerd-dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78.scope: Deactivated successfully. Apr 28 00:16:18.242277 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78-rootfs.mount: Deactivated successfully. Apr 28 00:16:18.252971 kubelet[2581]: I0428 00:16:18.251860 2581 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Apr 28 00:16:18.317744 containerd[1481]: time="2026-04-28T00:16:18.317664295Z" level=info msg="shim disconnected" id=dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78 namespace=k8s.io Apr 28 00:16:18.318015 containerd[1481]: time="2026-04-28T00:16:18.317997137Z" level=warning msg="cleaning up after shim disconnected" id=dc1019c2cc3d4b8511e68fd3a292e136bd5f05fcefaaff8fe7ca2d2ff6a62f78 namespace=k8s.io Apr 28 00:16:18.318301 containerd[1481]: time="2026-04-28T00:16:18.318119778Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 28 00:16:18.339754 systemd[1]: Created slice kubepods-besteffort-pod170ec0ef_bd09_4146_880c_4476c220befb.slice - libcontainer container kubepods-besteffort-pod170ec0ef_bd09_4146_880c_4476c220befb.slice. Apr 28 00:16:18.352169 systemd[1]: Created slice kubepods-burstable-poda74e08b1_013f_45ee_bd93_5b8aae9d758d.slice - libcontainer container kubepods-burstable-poda74e08b1_013f_45ee_bd93_5b8aae9d758d.slice. Apr 28 00:16:18.368293 systemd[1]: Created slice kubepods-besteffort-podd1762a6f_fcdd_4c15_ab45_2f2e776b2b7b.slice - libcontainer container kubepods-besteffort-podd1762a6f_fcdd_4c15_ab45_2f2e776b2b7b.slice. Apr 28 00:16:18.383026 systemd[1]: Created slice kubepods-besteffort-podf4111884_15d4_472c_b571_63308fbf58dd.slice - libcontainer container kubepods-besteffort-podf4111884_15d4_472c_b571_63308fbf58dd.slice. Apr 28 00:16:18.392039 systemd[1]: Created slice kubepods-besteffort-pod9d521bc8_2ea7_4994_aa05_4525d03b6763.slice - libcontainer container kubepods-besteffort-pod9d521bc8_2ea7_4994_aa05_4525d03b6763.slice. Apr 28 00:16:18.399924 systemd[1]: Created slice kubepods-besteffort-podc8d091df_86b9_498f_8893_12fb75b0c30e.slice - libcontainer container kubepods-besteffort-podc8d091df_86b9_498f_8893_12fb75b0c30e.slice. Apr 28 00:16:18.410435 systemd[1]: Created slice kubepods-burstable-pod90d5e942_3524_4477_a485_3644be314c39.slice - libcontainer container kubepods-burstable-pod90d5e942_3524_4477_a485_3644be314c39.slice. Apr 28 00:16:18.416470 kubelet[2581]: I0428 00:16:18.415899 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g97w9\" (UniqueName: \"kubernetes.io/projected/d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b-kube-api-access-g97w9\") pod \"calico-kube-controllers-8f7677868-s4r9s\" (UID: \"d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b\") " pod="calico-system/calico-kube-controllers-8f7677868-s4r9s" Apr 28 00:16:18.416470 kubelet[2581]: I0428 00:16:18.415946 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/170ec0ef-bd09-4146-880c-4476c220befb-whisker-backend-key-pair\") pod \"whisker-7486974bdf-9jnwq\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " pod="calico-system/whisker-7486974bdf-9jnwq" Apr 28 00:16:18.416470 kubelet[2581]: I0428 00:16:18.415966 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29rfg\" (UniqueName: \"kubernetes.io/projected/90d5e942-3524-4477-a485-3644be314c39-kube-api-access-29rfg\") pod \"coredns-674b8bbfcf-mqzhw\" (UID: \"90d5e942-3524-4477-a485-3644be314c39\") " pod="kube-system/coredns-674b8bbfcf-mqzhw" Apr 28 00:16:18.416470 kubelet[2581]: I0428 00:16:18.415988 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9cgn\" (UniqueName: \"kubernetes.io/projected/c8d091df-86b9-498f-8893-12fb75b0c30e-kube-api-access-b9cgn\") pod \"goldmane-57885fdd4c-glmtw\" (UID: \"c8d091df-86b9-498f-8893-12fb75b0c30e\") " pod="calico-system/goldmane-57885fdd4c-glmtw" Apr 28 00:16:18.416470 kubelet[2581]: I0428 00:16:18.416006 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90d5e942-3524-4477-a485-3644be314c39-config-volume\") pod \"coredns-674b8bbfcf-mqzhw\" (UID: \"90d5e942-3524-4477-a485-3644be314c39\") " pod="kube-system/coredns-674b8bbfcf-mqzhw" Apr 28 00:16:18.416935 kubelet[2581]: I0428 00:16:18.416025 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vzwt\" (UniqueName: \"kubernetes.io/projected/a74e08b1-013f-45ee-bd93-5b8aae9d758d-kube-api-access-2vzwt\") pod \"coredns-674b8bbfcf-zws2j\" (UID: \"a74e08b1-013f-45ee-bd93-5b8aae9d758d\") " pod="kube-system/coredns-674b8bbfcf-zws2j" Apr 28 00:16:18.416935 kubelet[2581]: I0428 00:16:18.416043 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c9lz\" (UniqueName: \"kubernetes.io/projected/f4111884-15d4-472c-b571-63308fbf58dd-kube-api-access-7c9lz\") pod \"calico-apiserver-674b8b8f7c-xw46n\" (UID: \"f4111884-15d4-472c-b571-63308fbf58dd\") " pod="calico-system/calico-apiserver-674b8b8f7c-xw46n" Apr 28 00:16:18.416935 kubelet[2581]: I0428 00:16:18.416059 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8d091df-86b9-498f-8893-12fb75b0c30e-config\") pod \"goldmane-57885fdd4c-glmtw\" (UID: \"c8d091df-86b9-498f-8893-12fb75b0c30e\") " pod="calico-system/goldmane-57885fdd4c-glmtw" Apr 28 00:16:18.416935 kubelet[2581]: I0428 00:16:18.416078 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/c8d091df-86b9-498f-8893-12fb75b0c30e-goldmane-key-pair\") pod \"goldmane-57885fdd4c-glmtw\" (UID: \"c8d091df-86b9-498f-8893-12fb75b0c30e\") " pod="calico-system/goldmane-57885fdd4c-glmtw" Apr 28 00:16:18.416935 kubelet[2581]: I0428 00:16:18.416101 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a74e08b1-013f-45ee-bd93-5b8aae9d758d-config-volume\") pod \"coredns-674b8bbfcf-zws2j\" (UID: \"a74e08b1-013f-45ee-bd93-5b8aae9d758d\") " pod="kube-system/coredns-674b8bbfcf-zws2j" Apr 28 00:16:18.417115 kubelet[2581]: I0428 00:16:18.416118 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f4111884-15d4-472c-b571-63308fbf58dd-calico-apiserver-certs\") pod \"calico-apiserver-674b8b8f7c-xw46n\" (UID: \"f4111884-15d4-472c-b571-63308fbf58dd\") " pod="calico-system/calico-apiserver-674b8b8f7c-xw46n" Apr 28 00:16:18.417115 kubelet[2581]: I0428 00:16:18.416146 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/9d521bc8-2ea7-4994-aa05-4525d03b6763-calico-apiserver-certs\") pod \"calico-apiserver-674b8b8f7c-jhxz2\" (UID: \"9d521bc8-2ea7-4994-aa05-4525d03b6763\") " pod="calico-system/calico-apiserver-674b8b8f7c-jhxz2" Apr 28 00:16:18.417115 kubelet[2581]: I0428 00:16:18.416164 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-whisker-ca-bundle\") pod \"whisker-7486974bdf-9jnwq\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " pod="calico-system/whisker-7486974bdf-9jnwq" Apr 28 00:16:18.417115 kubelet[2581]: I0428 00:16:18.416181 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-756xl\" (UniqueName: \"kubernetes.io/projected/170ec0ef-bd09-4146-880c-4476c220befb-kube-api-access-756xl\") pod \"whisker-7486974bdf-9jnwq\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " pod="calico-system/whisker-7486974bdf-9jnwq" Apr 28 00:16:18.417115 kubelet[2581]: I0428 00:16:18.416202 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs865\" (UniqueName: \"kubernetes.io/projected/9d521bc8-2ea7-4994-aa05-4525d03b6763-kube-api-access-xs865\") pod \"calico-apiserver-674b8b8f7c-jhxz2\" (UID: \"9d521bc8-2ea7-4994-aa05-4525d03b6763\") " pod="calico-system/calico-apiserver-674b8b8f7c-jhxz2" Apr 28 00:16:18.417353 kubelet[2581]: I0428 00:16:18.416228 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8d091df-86b9-498f-8893-12fb75b0c30e-goldmane-ca-bundle\") pod \"goldmane-57885fdd4c-glmtw\" (UID: \"c8d091df-86b9-498f-8893-12fb75b0c30e\") " pod="calico-system/goldmane-57885fdd4c-glmtw" Apr 28 00:16:18.417353 kubelet[2581]: I0428 00:16:18.416247 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b-tigera-ca-bundle\") pod \"calico-kube-controllers-8f7677868-s4r9s\" (UID: \"d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b\") " pod="calico-system/calico-kube-controllers-8f7677868-s4r9s" Apr 28 00:16:18.417353 kubelet[2581]: I0428 00:16:18.416265 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-nginx-config\") pod \"whisker-7486974bdf-9jnwq\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " pod="calico-system/whisker-7486974bdf-9jnwq" Apr 28 00:16:18.650029 containerd[1481]: time="2026-04-28T00:16:18.649864812Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7486974bdf-9jnwq,Uid:170ec0ef-bd09-4146-880c-4476c220befb,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:18.669709 containerd[1481]: time="2026-04-28T00:16:18.668468842Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-zws2j,Uid:a74e08b1-013f-45ee-bd93-5b8aae9d758d,Namespace:kube-system,Attempt:0,}" Apr 28 00:16:18.700707 containerd[1481]: time="2026-04-28T00:16:18.700422430Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-jhxz2,Uid:9d521bc8-2ea7-4994-aa05-4525d03b6763,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:18.701258 containerd[1481]: time="2026-04-28T00:16:18.701214315Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-8f7677868-s4r9s,Uid:d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:18.701468 containerd[1481]: time="2026-04-28T00:16:18.701439076Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-xw46n,Uid:f4111884-15d4-472c-b571-63308fbf58dd,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:18.708509 containerd[1481]: time="2026-04-28T00:16:18.708470718Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-57885fdd4c-glmtw,Uid:c8d091df-86b9-498f-8893-12fb75b0c30e,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:18.714565 containerd[1481]: time="2026-04-28T00:16:18.714275032Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-mqzhw,Uid:90d5e942-3524-4477-a485-3644be314c39,Namespace:kube-system,Attempt:0,}" Apr 28 00:16:18.787231 systemd[1]: Created slice kubepods-besteffort-pod7335aad5_4b0b_4297_8eed_20f9df1b81e5.slice - libcontainer container kubepods-besteffort-pod7335aad5_4b0b_4297_8eed_20f9df1b81e5.slice. Apr 28 00:16:18.799247 containerd[1481]: time="2026-04-28T00:16:18.798922491Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dcppv,Uid:7335aad5-4b0b-4297-8eed-20f9df1b81e5,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:18.862969 containerd[1481]: time="2026-04-28T00:16:18.862873187Z" level=error msg="Failed to destroy network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.866271 containerd[1481]: time="2026-04-28T00:16:18.865969806Z" level=error msg="encountered an error cleaning up failed sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.867378 containerd[1481]: time="2026-04-28T00:16:18.867325334Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7486974bdf-9jnwq,Uid:170ec0ef-bd09-4146-880c-4476c220befb,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.869307 kubelet[2581]: E0428 00:16:18.869140 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.869307 kubelet[2581]: E0428 00:16:18.869218 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7486974bdf-9jnwq" Apr 28 00:16:18.869307 kubelet[2581]: E0428 00:16:18.869237 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7486974bdf-9jnwq" Apr 28 00:16:18.870441 kubelet[2581]: E0428 00:16:18.869662 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7486974bdf-9jnwq_calico-system(170ec0ef-bd09-4146-880c-4476c220befb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7486974bdf-9jnwq_calico-system(170ec0ef-bd09-4146-880c-4476c220befb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7486974bdf-9jnwq" podUID="170ec0ef-bd09-4146-880c-4476c220befb" Apr 28 00:16:18.887105 containerd[1481]: time="2026-04-28T00:16:18.886969289Z" level=error msg="Failed to destroy network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.887427 containerd[1481]: time="2026-04-28T00:16:18.887392812Z" level=error msg="encountered an error cleaning up failed sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.887485 containerd[1481]: time="2026-04-28T00:16:18.887455292Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-zws2j,Uid:a74e08b1-013f-45ee-bd93-5b8aae9d758d,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.887742 kubelet[2581]: E0428 00:16:18.887661 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:18.887819 kubelet[2581]: E0428 00:16:18.887762 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-zws2j" Apr 28 00:16:18.887819 kubelet[2581]: E0428 00:16:18.887793 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-zws2j" Apr 28 00:16:18.887925 kubelet[2581]: E0428 00:16:18.887856 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-zws2j_kube-system(a74e08b1-013f-45ee-bd93-5b8aae9d758d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-zws2j_kube-system(a74e08b1-013f-45ee-bd93-5b8aae9d758d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-zws2j" podUID="a74e08b1-013f-45ee-bd93-5b8aae9d758d" Apr 28 00:16:18.921504 kubelet[2581]: I0428 00:16:18.921007 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:18.923600 containerd[1481]: time="2026-04-28T00:16:18.923256583Z" level=info msg="StopPodSandbox for \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\"" Apr 28 00:16:18.924659 containerd[1481]: time="2026-04-28T00:16:18.924300349Z" level=info msg="Ensure that sandbox cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4 in task-service has been cleanup successfully" Apr 28 00:16:18.926899 kubelet[2581]: I0428 00:16:18.926800 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:18.930213 containerd[1481]: time="2026-04-28T00:16:18.930156464Z" level=info msg="StopPodSandbox for \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\"" Apr 28 00:16:18.933052 containerd[1481]: time="2026-04-28T00:16:18.930654427Z" level=info msg="Ensure that sandbox 1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7 in task-service has been cleanup successfully" Apr 28 00:16:18.995749 containerd[1481]: time="2026-04-28T00:16:18.995547089Z" level=error msg="Failed to destroy network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.004703 containerd[1481]: time="2026-04-28T00:16:19.004586341Z" level=error msg="encountered an error cleaning up failed sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.004821 containerd[1481]: time="2026-04-28T00:16:19.004699542Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-mqzhw,Uid:90d5e942-3524-4477-a485-3644be314c39,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.005022 kubelet[2581]: E0428 00:16:19.004884 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.005022 kubelet[2581]: E0428 00:16:19.004943 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-mqzhw" Apr 28 00:16:19.005022 kubelet[2581]: E0428 00:16:19.004964 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-mqzhw" Apr 28 00:16:19.005462 kubelet[2581]: E0428 00:16:19.005014 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-mqzhw_kube-system(90d5e942-3524-4477-a485-3644be314c39)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-mqzhw_kube-system(90d5e942-3524-4477-a485-3644be314c39)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-mqzhw" podUID="90d5e942-3524-4477-a485-3644be314c39" Apr 28 00:16:19.089300 containerd[1481]: time="2026-04-28T00:16:19.089225168Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Apr 28 00:16:19.093377 containerd[1481]: time="2026-04-28T00:16:19.092964269Z" level=error msg="Failed to destroy network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.096346 containerd[1481]: time="2026-04-28T00:16:19.096280527Z" level=error msg="encountered an error cleaning up failed sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.096471 containerd[1481]: time="2026-04-28T00:16:19.096359008Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dcppv,Uid:7335aad5-4b0b-4297-8eed-20f9df1b81e5,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.098744 kubelet[2581]: E0428 00:16:19.098702 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.099089 kubelet[2581]: E0428 00:16:19.098992 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-dcppv" Apr 28 00:16:19.099089 kubelet[2581]: E0428 00:16:19.099054 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-dcppv" Apr 28 00:16:19.099295 kubelet[2581]: E0428 00:16:19.099268 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-dcppv_calico-system(7335aad5-4b0b-4297-8eed-20f9df1b81e5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-dcppv_calico-system(7335aad5-4b0b-4297-8eed-20f9df1b81e5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-dcppv" podUID="7335aad5-4b0b-4297-8eed-20f9df1b81e5" Apr 28 00:16:19.110303 containerd[1481]: time="2026-04-28T00:16:19.110255765Z" level=error msg="Failed to destroy network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.111036 containerd[1481]: time="2026-04-28T00:16:19.110802128Z" level=error msg="encountered an error cleaning up failed sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.111036 containerd[1481]: time="2026-04-28T00:16:19.110856528Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-jhxz2,Uid:9d521bc8-2ea7-4994-aa05-4525d03b6763,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.111199 kubelet[2581]: E0428 00:16:19.111066 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.111199 kubelet[2581]: E0428 00:16:19.111145 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-674b8b8f7c-jhxz2" Apr 28 00:16:19.111199 kubelet[2581]: E0428 00:16:19.111168 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-674b8b8f7c-jhxz2" Apr 28 00:16:19.111333 kubelet[2581]: E0428 00:16:19.111226 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-674b8b8f7c-jhxz2_calico-system(9d521bc8-2ea7-4994-aa05-4525d03b6763)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-674b8b8f7c-jhxz2_calico-system(9d521bc8-2ea7-4994-aa05-4525d03b6763)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-674b8b8f7c-jhxz2" podUID="9d521bc8-2ea7-4994-aa05-4525d03b6763" Apr 28 00:16:19.120216 containerd[1481]: time="2026-04-28T00:16:19.120054059Z" level=error msg="StopPodSandbox for \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\" failed" error="failed to destroy network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.120319 kubelet[2581]: E0428 00:16:19.120286 2581 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:19.120381 kubelet[2581]: E0428 00:16:19.120341 2581 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4"} Apr 28 00:16:19.120407 kubelet[2581]: E0428 00:16:19.120389 2581 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"a74e08b1-013f-45ee-bd93-5b8aae9d758d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 28 00:16:19.120491 kubelet[2581]: E0428 00:16:19.120412 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"a74e08b1-013f-45ee-bd93-5b8aae9d758d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-zws2j" podUID="a74e08b1-013f-45ee-bd93-5b8aae9d758d" Apr 28 00:16:19.143526 containerd[1481]: time="2026-04-28T00:16:19.142945425Z" level=info msg="CreateContainer within sandbox \"d312325a78a31436da4dc04fd82d43e7a0c933f25a3f4d7feaf31c64dc2a4fbb\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049\"" Apr 28 00:16:19.145451 containerd[1481]: time="2026-04-28T00:16:19.145383799Z" level=info msg="StartContainer for \"2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049\"" Apr 28 00:16:19.154143 containerd[1481]: time="2026-04-28T00:16:19.154078887Z" level=error msg="Failed to destroy network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.156215 containerd[1481]: time="2026-04-28T00:16:19.155964057Z" level=error msg="encountered an error cleaning up failed sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.156215 containerd[1481]: time="2026-04-28T00:16:19.156065898Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-xw46n,Uid:f4111884-15d4-472c-b571-63308fbf58dd,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.156847 kubelet[2581]: E0428 00:16:19.156516 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.156847 kubelet[2581]: E0428 00:16:19.156588 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-674b8b8f7c-xw46n" Apr 28 00:16:19.156847 kubelet[2581]: E0428 00:16:19.156607 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-674b8b8f7c-xw46n" Apr 28 00:16:19.157153 kubelet[2581]: E0428 00:16:19.156661 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-674b8b8f7c-xw46n_calico-system(f4111884-15d4-472c-b571-63308fbf58dd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-674b8b8f7c-xw46n_calico-system(f4111884-15d4-472c-b571-63308fbf58dd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-674b8b8f7c-xw46n" podUID="f4111884-15d4-472c-b571-63308fbf58dd" Apr 28 00:16:19.159931 containerd[1481]: time="2026-04-28T00:16:19.159884279Z" level=error msg="Failed to destroy network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.160752 containerd[1481]: time="2026-04-28T00:16:19.160273241Z" level=error msg="encountered an error cleaning up failed sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.160752 containerd[1481]: time="2026-04-28T00:16:19.160320281Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-8f7677868-s4r9s,Uid:d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.161395 kubelet[2581]: E0428 00:16:19.161344 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.161511 kubelet[2581]: E0428 00:16:19.161436 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-8f7677868-s4r9s" Apr 28 00:16:19.161558 kubelet[2581]: E0428 00:16:19.161461 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-8f7677868-s4r9s" Apr 28 00:16:19.162244 kubelet[2581]: E0428 00:16:19.161607 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-8f7677868-s4r9s_calico-system(d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-8f7677868-s4r9s_calico-system(d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-8f7677868-s4r9s" podUID="d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b" Apr 28 00:16:19.167464 containerd[1481]: time="2026-04-28T00:16:19.167393760Z" level=error msg="Failed to destroy network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.169082 containerd[1481]: time="2026-04-28T00:16:19.168924089Z" level=error msg="encountered an error cleaning up failed sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.169082 containerd[1481]: time="2026-04-28T00:16:19.168995569Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-57885fdd4c-glmtw,Uid:c8d091df-86b9-498f-8893-12fb75b0c30e,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.169267 kubelet[2581]: E0428 00:16:19.169193 2581 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.169267 kubelet[2581]: E0428 00:16:19.169240 2581 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-57885fdd4c-glmtw" Apr 28 00:16:19.169267 kubelet[2581]: E0428 00:16:19.169256 2581 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-57885fdd4c-glmtw" Apr 28 00:16:19.169362 kubelet[2581]: E0428 00:16:19.169300 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-57885fdd4c-glmtw_calico-system(c8d091df-86b9-498f-8893-12fb75b0c30e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-57885fdd4c-glmtw_calico-system(c8d091df-86b9-498f-8893-12fb75b0c30e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-57885fdd4c-glmtw" podUID="c8d091df-86b9-498f-8893-12fb75b0c30e" Apr 28 00:16:19.173055 containerd[1481]: time="2026-04-28T00:16:19.171787905Z" level=error msg="StopPodSandbox for \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\" failed" error="failed to destroy network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 28 00:16:19.173177 kubelet[2581]: E0428 00:16:19.172853 2581 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:19.173177 kubelet[2581]: E0428 00:16:19.172910 2581 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7"} Apr 28 00:16:19.173177 kubelet[2581]: E0428 00:16:19.172940 2581 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"170ec0ef-bd09-4146-880c-4476c220befb\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 28 00:16:19.173177 kubelet[2581]: E0428 00:16:19.172966 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"170ec0ef-bd09-4146-880c-4476c220befb\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7486974bdf-9jnwq" podUID="170ec0ef-bd09-4146-880c-4476c220befb" Apr 28 00:16:19.190903 systemd[1]: Started cri-containerd-2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049.scope - libcontainer container 2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049. Apr 28 00:16:19.231098 containerd[1481]: time="2026-04-28T00:16:19.231041952Z" level=info msg="StartContainer for \"2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049\" returns successfully" Apr 28 00:16:19.571187 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7-shm.mount: Deactivated successfully. Apr 28 00:16:20.042803 kubelet[2581]: I0428 00:16:20.042514 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:20.046933 containerd[1481]: time="2026-04-28T00:16:20.043261984Z" level=info msg="StopPodSandbox for \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\"" Apr 28 00:16:20.046933 containerd[1481]: time="2026-04-28T00:16:20.043429104Z" level=info msg="Ensure that sandbox 2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e in task-service has been cleanup successfully" Apr 28 00:16:20.053329 kubelet[2581]: I0428 00:16:20.053051 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:20.054600 containerd[1481]: time="2026-04-28T00:16:20.054385841Z" level=info msg="StopPodSandbox for \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\"" Apr 28 00:16:20.055571 containerd[1481]: time="2026-04-28T00:16:20.055216765Z" level=info msg="Ensure that sandbox a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803 in task-service has been cleanup successfully" Apr 28 00:16:20.055852 kubelet[2581]: I0428 00:16:20.055725 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:20.057163 containerd[1481]: time="2026-04-28T00:16:20.056662893Z" level=info msg="StopPodSandbox for \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\"" Apr 28 00:16:20.057163 containerd[1481]: time="2026-04-28T00:16:20.056910414Z" level=info msg="Ensure that sandbox b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf in task-service has been cleanup successfully" Apr 28 00:16:20.066329 kubelet[2581]: I0428 00:16:20.066206 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:20.066974 containerd[1481]: time="2026-04-28T00:16:20.066934306Z" level=info msg="StopPodSandbox for \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\"" Apr 28 00:16:20.067468 containerd[1481]: time="2026-04-28T00:16:20.067102587Z" level=info msg="Ensure that sandbox 8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc in task-service has been cleanup successfully" Apr 28 00:16:20.082470 kubelet[2581]: I0428 00:16:20.082157 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-x45hx" podStartSLOduration=3.8141114910000002 podStartE2EDuration="16.082102625s" podCreationTimestamp="2026-04-28 00:16:04 +0000 UTC" firstStartedPulling="2026-04-28 00:16:05.276793828 +0000 UTC m=+19.622888059" lastFinishedPulling="2026-04-28 00:16:17.544785042 +0000 UTC m=+31.890879193" observedRunningTime="2026-04-28 00:16:20.081873744 +0000 UTC m=+34.427968015" watchObservedRunningTime="2026-04-28 00:16:20.082102625 +0000 UTC m=+34.428196816" Apr 28 00:16:20.085878 kubelet[2581]: I0428 00:16:20.085825 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:20.088017 containerd[1481]: time="2026-04-28T00:16:20.087926455Z" level=info msg="StopPodSandbox for \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\"" Apr 28 00:16:20.088255 containerd[1481]: time="2026-04-28T00:16:20.088224296Z" level=info msg="Ensure that sandbox a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d in task-service has been cleanup successfully" Apr 28 00:16:20.093161 kubelet[2581]: I0428 00:16:20.093104 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:20.094611 containerd[1481]: time="2026-04-28T00:16:20.094555489Z" level=info msg="StopPodSandbox for \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\"" Apr 28 00:16:20.094780 containerd[1481]: time="2026-04-28T00:16:20.094758370Z" level=info msg="Ensure that sandbox 290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950 in task-service has been cleanup successfully" Apr 28 00:16:20.098852 containerd[1481]: time="2026-04-28T00:16:20.098811791Z" level=info msg="StopPodSandbox for \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\"" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.317 [INFO][3821] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.317 [INFO][3821] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" iface="eth0" netns="/var/run/netns/cni-398f0d7c-6135-a735-31b4-46eead213790" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.318 [INFO][3821] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" iface="eth0" netns="/var/run/netns/cni-398f0d7c-6135-a735-31b4-46eead213790" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.319 [INFO][3821] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" iface="eth0" netns="/var/run/netns/cni-398f0d7c-6135-a735-31b4-46eead213790" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.320 [INFO][3821] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.320 [INFO][3821] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.453 [INFO][3900] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.454 [INFO][3900] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.454 [INFO][3900] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.472 [WARNING][3900] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.472 [INFO][3900] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.478 [INFO][3900] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:20.485952 containerd[1481]: 2026-04-28 00:16:20.483 [INFO][3821] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:20.488482 containerd[1481]: time="2026-04-28T00:16:20.487736045Z" level=info msg="TearDown network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\" successfully" Apr 28 00:16:20.488482 containerd[1481]: time="2026-04-28T00:16:20.487789326Z" level=info msg="StopPodSandbox for \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\" returns successfully" Apr 28 00:16:20.490952 containerd[1481]: time="2026-04-28T00:16:20.490372939Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-xw46n,Uid:f4111884-15d4-472c-b571-63308fbf58dd,Namespace:calico-system,Attempt:1,}" Apr 28 00:16:20.491623 systemd[1]: run-netns-cni\x2d398f0d7c\x2d6135\x2da735\x2d31b4\x2d46eead213790.mount: Deactivated successfully. Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.323 [INFO][3807] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.323 [INFO][3807] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" iface="eth0" netns="/var/run/netns/cni-4d018fc7-1715-7dac-4f10-be5f766cd9fc" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.324 [INFO][3807] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" iface="eth0" netns="/var/run/netns/cni-4d018fc7-1715-7dac-4f10-be5f766cd9fc" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.325 [INFO][3807] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" iface="eth0" netns="/var/run/netns/cni-4d018fc7-1715-7dac-4f10-be5f766cd9fc" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.326 [INFO][3807] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.326 [INFO][3807] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.455 [INFO][3902] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.456 [INFO][3902] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.478 [INFO][3902] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.497 [WARNING][3902] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.497 [INFO][3902] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.501 [INFO][3902] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:20.536561 containerd[1481]: 2026-04-28 00:16:20.505 [INFO][3807] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:20.536960 containerd[1481]: time="2026-04-28T00:16:20.536868860Z" level=info msg="TearDown network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\" successfully" Apr 28 00:16:20.536960 containerd[1481]: time="2026-04-28T00:16:20.536898580Z" level=info msg="StopPodSandbox for \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\" returns successfully" Apr 28 00:16:20.541745 containerd[1481]: time="2026-04-28T00:16:20.541699485Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-mqzhw,Uid:90d5e942-3524-4477-a485-3644be314c39,Namespace:kube-system,Attempt:1,}" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.363 [INFO][3833] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.363 [INFO][3833] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" iface="eth0" netns="/var/run/netns/cni-40b4465e-dd8b-58eb-f82d-f81d0cf26361" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.364 [INFO][3833] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" iface="eth0" netns="/var/run/netns/cni-40b4465e-dd8b-58eb-f82d-f81d0cf26361" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.365 [INFO][3833] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" iface="eth0" netns="/var/run/netns/cni-40b4465e-dd8b-58eb-f82d-f81d0cf26361" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.365 [INFO][3833] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.365 [INFO][3833] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.456 [INFO][3915] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.457 [INFO][3915] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.501 [INFO][3915] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.523 [WARNING][3915] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.523 [INFO][3915] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.526 [INFO][3915] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:20.542365 containerd[1481]: 2026-04-28 00:16:20.536 [INFO][3833] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:20.542811 containerd[1481]: time="2026-04-28T00:16:20.542563409Z" level=info msg="TearDown network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\" successfully" Apr 28 00:16:20.542811 containerd[1481]: time="2026-04-28T00:16:20.542614410Z" level=info msg="StopPodSandbox for \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\" returns successfully" Apr 28 00:16:20.545273 containerd[1481]: time="2026-04-28T00:16:20.544979022Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-57885fdd4c-glmtw,Uid:c8d091df-86b9-498f-8893-12fb75b0c30e,Namespace:calico-system,Attempt:1,}" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.357 [INFO][3865] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.359 [INFO][3865] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" iface="eth0" netns="/var/run/netns/cni-688ac5da-1a66-8e24-8e11-b0e4bad48819" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.361 [INFO][3865] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" iface="eth0" netns="/var/run/netns/cni-688ac5da-1a66-8e24-8e11-b0e4bad48819" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.362 [INFO][3865] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" iface="eth0" netns="/var/run/netns/cni-688ac5da-1a66-8e24-8e11-b0e4bad48819" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.362 [INFO][3865] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.362 [INFO][3865] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.460 [INFO][3913] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.460 [INFO][3913] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.527 [INFO][3913] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.543 [WARNING][3913] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.543 [INFO][3913] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.546 [INFO][3913] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:20.555171 containerd[1481]: 2026-04-28 00:16:20.549 [INFO][3865] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:20.555954 containerd[1481]: time="2026-04-28T00:16:20.555609757Z" level=info msg="TearDown network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\" successfully" Apr 28 00:16:20.555954 containerd[1481]: time="2026-04-28T00:16:20.555886158Z" level=info msg="StopPodSandbox for \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\" returns successfully" Apr 28 00:16:20.557408 containerd[1481]: time="2026-04-28T00:16:20.557339126Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-jhxz2,Uid:9d521bc8-2ea7-4994-aa05-4525d03b6763,Namespace:calico-system,Attempt:1,}" Apr 28 00:16:20.561319 systemd[1]: run-netns-cni\x2d4d018fc7\x2d1715\x2d7dac\x2d4f10\x2dbe5f766cd9fc.mount: Deactivated successfully. Apr 28 00:16:20.561686 systemd[1]: run-netns-cni\x2d40b4465e\x2ddd8b\x2d58eb\x2df82d\x2df81d0cf26361.mount: Deactivated successfully. Apr 28 00:16:20.562417 systemd[1]: run-netns-cni\x2d688ac5da\x2d1a66\x2d8e24\x2d8e11\x2db0e4bad48819.mount: Deactivated successfully. Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.305 [INFO][3819] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.306 [INFO][3819] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" iface="eth0" netns="/var/run/netns/cni-c0a151dc-4bff-0526-05d2-3d3d6bd38388" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.307 [INFO][3819] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" iface="eth0" netns="/var/run/netns/cni-c0a151dc-4bff-0526-05d2-3d3d6bd38388" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.312 [INFO][3819] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" iface="eth0" netns="/var/run/netns/cni-c0a151dc-4bff-0526-05d2-3d3d6bd38388" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.313 [INFO][3819] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.313 [INFO][3819] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.454 [INFO][3899] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.461 [INFO][3899] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.546 [INFO][3899] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.576 [WARNING][3899] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.576 [INFO][3899] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.581 [INFO][3899] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:20.596961 containerd[1481]: 2026-04-28 00:16:20.587 [INFO][3819] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:20.601852 containerd[1481]: time="2026-04-28T00:16:20.601809236Z" level=info msg="TearDown network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\" successfully" Apr 28 00:16:20.603742 containerd[1481]: time="2026-04-28T00:16:20.602983282Z" level=info msg="StopPodSandbox for \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\" returns successfully" Apr 28 00:16:20.605720 systemd[1]: run-netns-cni\x2dc0a151dc\x2d4bff\x2d0526\x2d05d2\x2d3d3d6bd38388.mount: Deactivated successfully. Apr 28 00:16:20.606579 containerd[1481]: time="2026-04-28T00:16:20.605888657Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dcppv,Uid:7335aad5-4b0b-4297-8eed-20f9df1b81e5,Namespace:calico-system,Attempt:1,}" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.365 [INFO][3877] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.367 [INFO][3877] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" iface="eth0" netns="/var/run/netns/cni-aef20bf4-cef2-a5e5-e021-ec569e2c7142" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.368 [INFO][3877] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" iface="eth0" netns="/var/run/netns/cni-aef20bf4-cef2-a5e5-e021-ec569e2c7142" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.370 [INFO][3877] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" iface="eth0" netns="/var/run/netns/cni-aef20bf4-cef2-a5e5-e021-ec569e2c7142" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.370 [INFO][3877] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.370 [INFO][3877] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.477 [INFO][3918] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.477 [INFO][3918] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.582 [INFO][3918] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.624 [WARNING][3918] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.624 [INFO][3918] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.632 [INFO][3918] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:20.694550 containerd[1481]: 2026-04-28 00:16:20.638 [INFO][3877] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.369 [INFO][3866] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.369 [INFO][3866] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" iface="eth0" netns="/var/run/netns/cni-b1fe4489-e910-46db-b0e0-d8e18e16226a" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.369 [INFO][3866] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" iface="eth0" netns="/var/run/netns/cni-b1fe4489-e910-46db-b0e0-d8e18e16226a" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.372 [INFO][3866] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" iface="eth0" netns="/var/run/netns/cni-b1fe4489-e910-46db-b0e0-d8e18e16226a" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.372 [INFO][3866] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.372 [INFO][3866] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.477 [INFO][3922] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.478 [INFO][3922] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.632 [INFO][3922] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.662 [WARNING][3922] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.662 [INFO][3922] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.669 [INFO][3922] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:20.702700 containerd[1481]: 2026-04-28 00:16:20.674 [INFO][3866] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:20.717610 containerd[1481]: time="2026-04-28T00:16:20.717525875Z" level=info msg="TearDown network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\" successfully" Apr 28 00:16:20.717610 containerd[1481]: time="2026-04-28T00:16:20.717591996Z" level=info msg="StopPodSandbox for \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\" returns successfully" Apr 28 00:16:20.720230 containerd[1481]: time="2026-04-28T00:16:20.720174489Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-8f7677868-s4r9s,Uid:d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b,Namespace:calico-system,Attempt:1,}" Apr 28 00:16:20.723276 containerd[1481]: time="2026-04-28T00:16:20.722816663Z" level=info msg="TearDown network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\" successfully" Apr 28 00:16:20.723276 containerd[1481]: time="2026-04-28T00:16:20.722860823Z" level=info msg="StopPodSandbox for \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\" returns successfully" Apr 28 00:16:20.833193 kubelet[2581]: I0428 00:16:20.833034 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-nginx-config\") pod \"170ec0ef-bd09-4146-880c-4476c220befb\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " Apr 28 00:16:20.833744 kubelet[2581]: I0428 00:16:20.833727 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/170ec0ef-bd09-4146-880c-4476c220befb-whisker-backend-key-pair\") pod \"170ec0ef-bd09-4146-880c-4476c220befb\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " Apr 28 00:16:20.835980 kubelet[2581]: I0428 00:16:20.835952 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-whisker-ca-bundle\") pod \"170ec0ef-bd09-4146-880c-4476c220befb\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " Apr 28 00:16:20.839683 kubelet[2581]: I0428 00:16:20.837095 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-756xl\" (UniqueName: \"kubernetes.io/projected/170ec0ef-bd09-4146-880c-4476c220befb-kube-api-access-756xl\") pod \"170ec0ef-bd09-4146-880c-4476c220befb\" (UID: \"170ec0ef-bd09-4146-880c-4476c220befb\") " Apr 28 00:16:20.839683 kubelet[2581]: I0428 00:16:20.837282 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-nginx-config" (OuterVolumeSpecName: "nginx-config") pod "170ec0ef-bd09-4146-880c-4476c220befb" (UID: "170ec0ef-bd09-4146-880c-4476c220befb"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 28 00:16:20.844600 kubelet[2581]: I0428 00:16:20.843384 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "170ec0ef-bd09-4146-880c-4476c220befb" (UID: "170ec0ef-bd09-4146-880c-4476c220befb"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 28 00:16:20.849434 kubelet[2581]: I0428 00:16:20.849325 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/170ec0ef-bd09-4146-880c-4476c220befb-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "170ec0ef-bd09-4146-880c-4476c220befb" (UID: "170ec0ef-bd09-4146-880c-4476c220befb"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 28 00:16:20.858112 kubelet[2581]: I0428 00:16:20.858069 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/170ec0ef-bd09-4146-880c-4476c220befb-kube-api-access-756xl" (OuterVolumeSpecName: "kube-api-access-756xl") pod "170ec0ef-bd09-4146-880c-4476c220befb" (UID: "170ec0ef-bd09-4146-880c-4476c220befb"). InnerVolumeSpecName "kube-api-access-756xl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 28 00:16:20.937749 kubelet[2581]: I0428 00:16:20.937707 2581 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/170ec0ef-bd09-4146-880c-4476c220befb-whisker-backend-key-pair\") on node \"ci-4081-3-7-n-51c70c830d\" DevicePath \"\"" Apr 28 00:16:20.937749 kubelet[2581]: I0428 00:16:20.937742 2581 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-whisker-ca-bundle\") on node \"ci-4081-3-7-n-51c70c830d\" DevicePath \"\"" Apr 28 00:16:20.937749 kubelet[2581]: I0428 00:16:20.937753 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-756xl\" (UniqueName: \"kubernetes.io/projected/170ec0ef-bd09-4146-880c-4476c220befb-kube-api-access-756xl\") on node \"ci-4081-3-7-n-51c70c830d\" DevicePath \"\"" Apr 28 00:16:20.937749 kubelet[2581]: I0428 00:16:20.937762 2581 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/170ec0ef-bd09-4146-880c-4476c220befb-nginx-config\") on node \"ci-4081-3-7-n-51c70c830d\" DevicePath \"\"" Apr 28 00:16:21.075250 systemd-networkd[1371]: cali0eb5e55d91c: Link UP Apr 28 00:16:21.078000 systemd-networkd[1371]: cali0eb5e55d91c: Gained carrier Apr 28 00:16:21.108186 systemd[1]: Removed slice kubepods-besteffort-pod170ec0ef_bd09_4146_880c_4476c220befb.slice - libcontainer container kubepods-besteffort-pod170ec0ef_bd09_4146_880c_4476c220befb.slice. Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.579 [ERROR][3945] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.640 [INFO][3945] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0 calico-apiserver-674b8b8f7c- calico-system f4111884-15d4-472c-b571-63308fbf58dd 918 0 2026-04-28 00:16:03 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:674b8b8f7c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d calico-apiserver-674b8b8f7c-xw46n eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali0eb5e55d91c [] [] }} ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.642 [INFO][3945] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.890 [INFO][4015] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" HandleID="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.927 [INFO][4015] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" HandleID="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004f1d0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"calico-apiserver-674b8b8f7c-xw46n", "timestamp":"2026-04-28 00:16:20.890051089 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000189340)} Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.927 [INFO][4015] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.927 [INFO][4015] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.927 [INFO][4015] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.945 [INFO][4015] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.961 [INFO][4015] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.975 [INFO][4015] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.980 [INFO][4015] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.984 [INFO][4015] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.984 [INFO][4015] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:20.994 [INFO][4015] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:21.008 [INFO][4015] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:21.027 [INFO][4015] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.129/26] block=192.168.8.128/26 handle="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:21.030 [INFO][4015] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.129/26] handle="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:21.030 [INFO][4015] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:21.130011 containerd[1481]: 2026-04-28 00:16:21.030 [INFO][4015] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.129/26] IPv6=[] ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" HandleID="k8s-pod-network.1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:21.132265 containerd[1481]: 2026-04-28 00:16:21.040 [INFO][3945] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"f4111884-15d4-472c-b571-63308fbf58dd", ResourceVersion:"918", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"calico-apiserver-674b8b8f7c-xw46n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali0eb5e55d91c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.132265 containerd[1481]: 2026-04-28 00:16:21.041 [INFO][3945] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.129/32] ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:21.132265 containerd[1481]: 2026-04-28 00:16:21.042 [INFO][3945] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0eb5e55d91c ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:21.132265 containerd[1481]: 2026-04-28 00:16:21.076 [INFO][3945] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:21.132265 containerd[1481]: 2026-04-28 00:16:21.080 [INFO][3945] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"f4111884-15d4-472c-b571-63308fbf58dd", ResourceVersion:"918", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e", Pod:"calico-apiserver-674b8b8f7c-xw46n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali0eb5e55d91c", MAC:"ce:b6:21:c4:40:c4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.132265 containerd[1481]: 2026-04-28 00:16:21.121 [INFO][3945] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-xw46n" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:21.242825 systemd-networkd[1371]: calid189784c50e: Link UP Apr 28 00:16:21.243178 systemd-networkd[1371]: calid189784c50e: Gained carrier Apr 28 00:16:21.271710 containerd[1481]: time="2026-04-28T00:16:21.269189125Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:21.271710 containerd[1481]: time="2026-04-28T00:16:21.269273166Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:21.271710 containerd[1481]: time="2026-04-28T00:16:21.269289686Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.271710 containerd[1481]: time="2026-04-28T00:16:21.269409366Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.272944 systemd[1]: Created slice kubepods-besteffort-podbd6025bd_4014_436d_aeea_02adcdff7a99.slice - libcontainer container kubepods-besteffort-podbd6025bd_4014_436d_aeea_02adcdff7a99.slice. Apr 28 00:16:21.316300 systemd[1]: Started cri-containerd-1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e.scope - libcontainer container 1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e. Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:20.753 [ERROR][4010] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:20.790 [INFO][4010] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0 calico-apiserver-674b8b8f7c- calico-system 9d521bc8-2ea7-4994-aa05-4525d03b6763 920 0 2026-04-28 00:16:03 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:674b8b8f7c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d calico-apiserver-674b8b8f7c-jhxz2 eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] calid189784c50e [] [] }} ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:20.790 [INFO][4010] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:20.978 [INFO][4089] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" HandleID="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.010 [INFO][4089] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" HandleID="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400026be90), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"calico-apiserver-674b8b8f7c-jhxz2", "timestamp":"2026-04-28 00:16:20.977999024 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400010d600)} Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.010 [INFO][4089] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.034 [INFO][4089] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.034 [INFO][4089] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.043 [INFO][4089] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.079 [INFO][4089] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.126 [INFO][4089] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.145 [INFO][4089] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.156 [INFO][4089] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.156 [INFO][4089] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.170 [INFO][4089] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795 Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.190 [INFO][4089] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.213 [INFO][4089] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.130/26] block=192.168.8.128/26 handle="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.214 [INFO][4089] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.130/26] handle="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.215 [INFO][4089] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:21.329903 containerd[1481]: 2026-04-28 00:16:21.216 [INFO][4089] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.130/26] IPv6=[] ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" HandleID="k8s-pod-network.7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:21.332214 containerd[1481]: 2026-04-28 00:16:21.226 [INFO][4010] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"9d521bc8-2ea7-4994-aa05-4525d03b6763", ResourceVersion:"920", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"calico-apiserver-674b8b8f7c-jhxz2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calid189784c50e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.332214 containerd[1481]: 2026-04-28 00:16:21.227 [INFO][4010] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.130/32] ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:21.332214 containerd[1481]: 2026-04-28 00:16:21.227 [INFO][4010] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid189784c50e ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:21.332214 containerd[1481]: 2026-04-28 00:16:21.241 [INFO][4010] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:21.332214 containerd[1481]: 2026-04-28 00:16:21.250 [INFO][4010] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"9d521bc8-2ea7-4994-aa05-4525d03b6763", ResourceVersion:"920", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795", Pod:"calico-apiserver-674b8b8f7c-jhxz2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calid189784c50e", MAC:"72:6c:d5:67:27:9a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.332214 containerd[1481]: 2026-04-28 00:16:21.318 [INFO][4010] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795" Namespace="calico-system" Pod="calico-apiserver-674b8b8f7c-jhxz2" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:21.342363 kubelet[2581]: I0428 00:16:21.341908 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd6025bd-4014-436d-aeea-02adcdff7a99-whisker-ca-bundle\") pod \"whisker-5d8cd77c89-ccd2x\" (UID: \"bd6025bd-4014-436d-aeea-02adcdff7a99\") " pod="calico-system/whisker-5d8cd77c89-ccd2x" Apr 28 00:16:21.344151 kubelet[2581]: I0428 00:16:21.343874 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/bd6025bd-4014-436d-aeea-02adcdff7a99-nginx-config\") pod \"whisker-5d8cd77c89-ccd2x\" (UID: \"bd6025bd-4014-436d-aeea-02adcdff7a99\") " pod="calico-system/whisker-5d8cd77c89-ccd2x" Apr 28 00:16:21.344151 kubelet[2581]: I0428 00:16:21.343919 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbshm\" (UniqueName: \"kubernetes.io/projected/bd6025bd-4014-436d-aeea-02adcdff7a99-kube-api-access-sbshm\") pod \"whisker-5d8cd77c89-ccd2x\" (UID: \"bd6025bd-4014-436d-aeea-02adcdff7a99\") " pod="calico-system/whisker-5d8cd77c89-ccd2x" Apr 28 00:16:21.344802 kubelet[2581]: I0428 00:16:21.344337 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/bd6025bd-4014-436d-aeea-02adcdff7a99-whisker-backend-key-pair\") pod \"whisker-5d8cd77c89-ccd2x\" (UID: \"bd6025bd-4014-436d-aeea-02adcdff7a99\") " pod="calico-system/whisker-5d8cd77c89-ccd2x" Apr 28 00:16:21.422751 containerd[1481]: time="2026-04-28T00:16:21.422474829Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:21.423588 containerd[1481]: time="2026-04-28T00:16:21.422560990Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:21.423588 containerd[1481]: time="2026-04-28T00:16:21.423154873Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.423588 containerd[1481]: time="2026-04-28T00:16:21.423289633Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.430397 systemd-networkd[1371]: cali0e4558a8691: Link UP Apr 28 00:16:21.433441 systemd-networkd[1371]: cali0e4558a8691: Gained carrier Apr 28 00:16:21.478947 containerd[1481]: time="2026-04-28T00:16:21.478656342Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-xw46n,Uid:f4111884-15d4-472c-b571-63308fbf58dd,Namespace:calico-system,Attempt:1,} returns sandbox id \"1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e\"" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:20.804 [ERROR][3959] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:20.883 [INFO][3959] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0 coredns-674b8bbfcf- kube-system 90d5e942-3524-4477-a485-3644be314c39 919 0 2026-04-28 00:15:50 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d coredns-674b8bbfcf-mqzhw eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali0e4558a8691 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:20.884 [INFO][3959] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.034 [INFO][4112] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" HandleID="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.062 [INFO][4112] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" HandleID="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ee170), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"coredns-674b8bbfcf-mqzhw", "timestamp":"2026-04-28 00:16:21.034536906 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000188580)} Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.063 [INFO][4112] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.215 [INFO][4112] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.215 [INFO][4112] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.322 [INFO][4112] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.355 [INFO][4112] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.366 [INFO][4112] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.373 [INFO][4112] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.377 [INFO][4112] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.377 [INFO][4112] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.381 [INFO][4112] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.389 [INFO][4112] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.402 [INFO][4112] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.131/26] block=192.168.8.128/26 handle="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.402 [INFO][4112] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.131/26] handle="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.404 [INFO][4112] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:21.488538 containerd[1481]: 2026-04-28 00:16:21.404 [INFO][4112] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.131/26] IPv6=[] ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" HandleID="k8s-pod-network.1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:21.489177 containerd[1481]: 2026-04-28 00:16:21.421 [INFO][3959] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"90d5e942-3524-4477-a485-3644be314c39", ResourceVersion:"919", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"coredns-674b8bbfcf-mqzhw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali0e4558a8691", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.489177 containerd[1481]: 2026-04-28 00:16:21.422 [INFO][3959] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.131/32] ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:21.489177 containerd[1481]: 2026-04-28 00:16:21.422 [INFO][3959] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0e4558a8691 ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:21.489177 containerd[1481]: 2026-04-28 00:16:21.428 [INFO][3959] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:21.489177 containerd[1481]: 2026-04-28 00:16:21.437 [INFO][3959] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"90d5e942-3524-4477-a485-3644be314c39", ResourceVersion:"919", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c", Pod:"coredns-674b8bbfcf-mqzhw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali0e4558a8691", MAC:"d6:a3:2e:6e:9a:4e", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.489177 containerd[1481]: 2026-04-28 00:16:21.483 [INFO][3959] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c" Namespace="kube-system" Pod="coredns-674b8bbfcf-mqzhw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:21.498779 systemd[1]: Started cri-containerd-7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795.scope - libcontainer container 7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795. Apr 28 00:16:21.505600 containerd[1481]: time="2026-04-28T00:16:21.504799389Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\"" Apr 28 00:16:21.547259 containerd[1481]: time="2026-04-28T00:16:21.542294971Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:21.547259 containerd[1481]: time="2026-04-28T00:16:21.542366052Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:21.547259 containerd[1481]: time="2026-04-28T00:16:21.542386812Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.547259 containerd[1481]: time="2026-04-28T00:16:21.542494972Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.553539 systemd-networkd[1371]: cali7413111db7d: Link UP Apr 28 00:16:21.553823 systemd-networkd[1371]: cali7413111db7d: Gained carrier Apr 28 00:16:21.577629 systemd[1]: run-netns-cni\x2daef20bf4\x2dcef2\x2da5e5\x2de021\x2dec569e2c7142.mount: Deactivated successfully. Apr 28 00:16:21.577749 systemd[1]: run-netns-cni\x2db1fe4489\x2de910\x2d46db\x2db0e0\x2dd8e18e16226a.mount: Deactivated successfully. Apr 28 00:16:21.578538 systemd[1]: var-lib-kubelet-pods-170ec0ef\x2dbd09\x2d4146\x2d880c\x2d4476c220befb-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d756xl.mount: Deactivated successfully. Apr 28 00:16:21.578637 systemd[1]: var-lib-kubelet-pods-170ec0ef\x2dbd09\x2d4146\x2d880c\x2d4476c220befb-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Apr 28 00:16:21.585588 containerd[1481]: time="2026-04-28T00:16:21.585547621Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d8cd77c89-ccd2x,Uid:bd6025bd-4014-436d-aeea-02adcdff7a99,Namespace:calico-system,Attempt:0,}" Apr 28 00:16:21.607904 systemd[1]: Started cri-containerd-1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c.scope - libcontainer container 1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c. Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:20.799 [ERROR][3960] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:20.891 [INFO][3960] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0 goldmane-57885fdd4c- calico-system c8d091df-86b9-498f-8893-12fb75b0c30e 922 0 2026-04-28 00:16:04 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:57885fdd4c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d goldmane-57885fdd4c-glmtw eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali7413111db7d [] [] }} ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:20.891 [INFO][3960] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.077 [INFO][4102] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" HandleID="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.124 [INFO][4102] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" HandleID="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004f640), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"goldmane-57885fdd4c-glmtw", "timestamp":"2026-04-28 00:16:21.077803716 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40000e4160)} Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.124 [INFO][4102] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.404 [INFO][4102] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.404 [INFO][4102] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.410 [INFO][4102] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.473 [INFO][4102] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.492 [INFO][4102] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.497 [INFO][4102] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.504 [INFO][4102] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.504 [INFO][4102] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.508 [INFO][4102] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.515 [INFO][4102] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.529 [INFO][4102] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.132/26] block=192.168.8.128/26 handle="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.529 [INFO][4102] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.132/26] handle="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.529 [INFO][4102] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:21.631996 containerd[1481]: 2026-04-28 00:16:21.529 [INFO][4102] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.132/26] IPv6=[] ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" HandleID="k8s-pod-network.679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:21.632558 containerd[1481]: 2026-04-28 00:16:21.535 [INFO][3960] cni-plugin/k8s.go 418: Populated endpoint ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0", GenerateName:"goldmane-57885fdd4c-", Namespace:"calico-system", SelfLink:"", UID:"c8d091df-86b9-498f-8893-12fb75b0c30e", ResourceVersion:"922", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"57885fdd4c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"goldmane-57885fdd4c-glmtw", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.8.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7413111db7d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.632558 containerd[1481]: 2026-04-28 00:16:21.536 [INFO][3960] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.132/32] ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:21.632558 containerd[1481]: 2026-04-28 00:16:21.536 [INFO][3960] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7413111db7d ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:21.632558 containerd[1481]: 2026-04-28 00:16:21.553 [INFO][3960] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:21.632558 containerd[1481]: 2026-04-28 00:16:21.566 [INFO][3960] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0", GenerateName:"goldmane-57885fdd4c-", Namespace:"calico-system", SelfLink:"", UID:"c8d091df-86b9-498f-8893-12fb75b0c30e", ResourceVersion:"922", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"57885fdd4c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b", Pod:"goldmane-57885fdd4c-glmtw", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.8.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7413111db7d", MAC:"1a:3e:fa:2c:3c:46", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.632558 containerd[1481]: 2026-04-28 00:16:21.622 [INFO][3960] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b" Namespace="calico-system" Pod="goldmane-57885fdd4c-glmtw" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:21.720490 containerd[1481]: time="2026-04-28T00:16:21.712897519Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:21.720490 containerd[1481]: time="2026-04-28T00:16:21.720297155Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:21.721767 containerd[1481]: time="2026-04-28T00:16:21.720323716Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.722703 containerd[1481]: time="2026-04-28T00:16:21.721973324Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.761633 systemd-networkd[1371]: calib31e86a8995: Link UP Apr 28 00:16:21.770558 systemd-networkd[1371]: calib31e86a8995: Gained carrier Apr 28 00:16:21.785690 kubelet[2581]: I0428 00:16:21.785348 2581 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="170ec0ef-bd09-4146-880c-4476c220befb" path="/var/lib/kubelet/pods/170ec0ef-bd09-4146-880c-4476c220befb/volumes" Apr 28 00:16:21.795979 containerd[1481]: time="2026-04-28T00:16:21.794467156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-mqzhw,Uid:90d5e942-3524-4477-a485-3644be314c39,Namespace:kube-system,Attempt:1,} returns sandbox id \"1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c\"" Apr 28 00:16:21.813522 containerd[1481]: time="2026-04-28T00:16:21.810990756Z" level=info msg="CreateContainer within sandbox \"1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 28 00:16:21.817070 systemd[1]: Started cri-containerd-679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b.scope - libcontainer container 679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b. Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:20.925 [ERROR][4036] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:20.965 [INFO][4036] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0 csi-node-driver- calico-system 7335aad5-4b0b-4297-8eed-20f9df1b81e5 917 0 2026-04-28 00:16:04 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:74865c565 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d csi-node-driver-dcppv eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calib31e86a8995 [] [] }} ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:20.965 [INFO][4036] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.098 [INFO][4136] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" HandleID="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.154 [INFO][4136] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" HandleID="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000127eb0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"csi-node-driver-dcppv", "timestamp":"2026-04-28 00:16:21.098818298 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000189080)} Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.155 [INFO][4136] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.529 [INFO][4136] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.529 [INFO][4136] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.534 [INFO][4136] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.591 [INFO][4136] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.638 [INFO][4136] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.645 [INFO][4136] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.659 [INFO][4136] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.659 [INFO][4136] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.672 [INFO][4136] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05 Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.700 [INFO][4136] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.739 [INFO][4136] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.133/26] block=192.168.8.128/26 handle="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.739 [INFO][4136] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.133/26] handle="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.739 [INFO][4136] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:21.834171 containerd[1481]: 2026-04-28 00:16:21.739 [INFO][4136] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.133/26] IPv6=[] ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" HandleID="k8s-pod-network.0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:21.834731 containerd[1481]: 2026-04-28 00:16:21.746 [INFO][4036] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7335aad5-4b0b-4297-8eed-20f9df1b81e5", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"74865c565", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"csi-node-driver-dcppv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.8.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calib31e86a8995", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.834731 containerd[1481]: 2026-04-28 00:16:21.749 [INFO][4036] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.133/32] ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:21.834731 containerd[1481]: 2026-04-28 00:16:21.749 [INFO][4036] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib31e86a8995 ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:21.834731 containerd[1481]: 2026-04-28 00:16:21.775 [INFO][4036] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:21.834731 containerd[1481]: 2026-04-28 00:16:21.789 [INFO][4036] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7335aad5-4b0b-4297-8eed-20f9df1b81e5", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"74865c565", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05", Pod:"csi-node-driver-dcppv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.8.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calib31e86a8995", MAC:"7a:55:78:21:40:91", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.834731 containerd[1481]: 2026-04-28 00:16:21.828 [INFO][4036] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05" Namespace="calico-system" Pod="csi-node-driver-dcppv" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:21.839899 containerd[1481]: time="2026-04-28T00:16:21.839841616Z" level=info msg="CreateContainer within sandbox \"1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"47f808ac971e9f998a712229cbd425c1c11e3b55f3c1e42777add8454d51e723\"" Apr 28 00:16:21.840776 containerd[1481]: time="2026-04-28T00:16:21.840589419Z" level=info msg="StartContainer for \"47f808ac971e9f998a712229cbd425c1c11e3b55f3c1e42777add8454d51e723\"" Apr 28 00:16:21.900800 containerd[1481]: time="2026-04-28T00:16:21.899365025Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:21.900800 containerd[1481]: time="2026-04-28T00:16:21.899427785Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:21.900800 containerd[1481]: time="2026-04-28T00:16:21.899442665Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.900800 containerd[1481]: time="2026-04-28T00:16:21.899529906Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:21.924465 systemd-networkd[1371]: cali2d5552d3884: Link UP Apr 28 00:16:21.926482 systemd-networkd[1371]: cali2d5552d3884: Gained carrier Apr 28 00:16:21.931025 systemd[1]: Started cri-containerd-47f808ac971e9f998a712229cbd425c1c11e3b55f3c1e42777add8454d51e723.scope - libcontainer container 47f808ac971e9f998a712229cbd425c1c11e3b55f3c1e42777add8454d51e723. Apr 28 00:16:21.963962 systemd[1]: Started cri-containerd-0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05.scope - libcontainer container 0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05. Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:20.981 [ERROR][4054] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.043 [INFO][4054] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0 calico-kube-controllers-8f7677868- calico-system d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b 921 0 2026-04-28 00:16:05 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:8f7677868 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d calico-kube-controllers-8f7677868-s4r9s eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali2d5552d3884 [] [] }} ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.044 [INFO][4054] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.225 [INFO][4154] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" HandleID="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.299 [INFO][4154] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" HandleID="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400031bea0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"calico-kube-controllers-8f7677868-s4r9s", "timestamp":"2026-04-28 00:16:21.225045671 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40001bcf20)} Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.299 [INFO][4154] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.739 [INFO][4154] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.742 [INFO][4154] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.749 [INFO][4154] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.768 [INFO][4154] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.789 [INFO][4154] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.822 [INFO][4154] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.836 [INFO][4154] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.836 [INFO][4154] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.850 [INFO][4154] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16 Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.886 [INFO][4154] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.910 [INFO][4154] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.134/26] block=192.168.8.128/26 handle="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.910 [INFO][4154] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.134/26] handle="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.910 [INFO][4154] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:21.978070 containerd[1481]: 2026-04-28 00:16:21.910 [INFO][4154] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.134/26] IPv6=[] ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" HandleID="k8s-pod-network.480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:21.979381 containerd[1481]: 2026-04-28 00:16:21.917 [INFO][4054] cni-plugin/k8s.go 418: Populated endpoint ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0", GenerateName:"calico-kube-controllers-8f7677868-", Namespace:"calico-system", SelfLink:"", UID:"d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b", ResourceVersion:"921", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"8f7677868", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"calico-kube-controllers-8f7677868-s4r9s", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.8.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2d5552d3884", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.979381 containerd[1481]: 2026-04-28 00:16:21.918 [INFO][4054] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.134/32] ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:21.979381 containerd[1481]: 2026-04-28 00:16:21.918 [INFO][4054] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2d5552d3884 ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:21.979381 containerd[1481]: 2026-04-28 00:16:21.938 [INFO][4054] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:21.979381 containerd[1481]: 2026-04-28 00:16:21.946 [INFO][4054] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0", GenerateName:"calico-kube-controllers-8f7677868-", Namespace:"calico-system", SelfLink:"", UID:"d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b", ResourceVersion:"921", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"8f7677868", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16", Pod:"calico-kube-controllers-8f7677868-s4r9s", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.8.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2d5552d3884", MAC:"ae:3f:ab:c9:b2:0d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:21.979381 containerd[1481]: 2026-04-28 00:16:21.975 [INFO][4054] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16" Namespace="calico-system" Pod="calico-kube-controllers-8f7677868-s4r9s" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:22.015746 containerd[1481]: time="2026-04-28T00:16:22.014447539Z" level=info msg="StartContainer for \"47f808ac971e9f998a712229cbd425c1c11e3b55f3c1e42777add8454d51e723\" returns successfully" Apr 28 00:16:22.028607 containerd[1481]: time="2026-04-28T00:16:22.028450923Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:22.028607 containerd[1481]: time="2026-04-28T00:16:22.028541283Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:22.028607 containerd[1481]: time="2026-04-28T00:16:22.028553004Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:22.029701 containerd[1481]: time="2026-04-28T00:16:22.028896765Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:22.072007 systemd[1]: Started cri-containerd-480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16.scope - libcontainer container 480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16. Apr 28 00:16:22.103073 systemd-networkd[1371]: cali28ee2b32562: Link UP Apr 28 00:16:22.107247 systemd-networkd[1371]: cali28ee2b32562: Gained carrier Apr 28 00:16:22.139835 containerd[1481]: time="2026-04-28T00:16:22.139739230Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-674b8b8f7c-jhxz2,Uid:9d521bc8-2ea7-4994-aa05-4525d03b6763,Namespace:calico-system,Attempt:1,} returns sandbox id \"7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795\"" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.712 [ERROR][4314] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.808 [INFO][4314] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0 whisker-5d8cd77c89- calico-system bd6025bd-4014-436d-aeea-02adcdff7a99 945 0 2026-04-28 00:16:21 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:5d8cd77c89 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d whisker-5d8cd77c89-ccd2x eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali28ee2b32562 [] [] }} ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.810 [INFO][4314] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.970 [INFO][4384] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" HandleID="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.992 [INFO][4384] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" HandleID="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400026b1b0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"whisker-5d8cd77c89-ccd2x", "timestamp":"2026-04-28 00:16:21.970003008 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400010f080)} Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.993 [INFO][4384] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.993 [INFO][4384] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:21.994 [INFO][4384] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.000 [INFO][4384] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.015 [INFO][4384] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.026 [INFO][4384] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.032 [INFO][4384] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.040 [INFO][4384] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.040 [INFO][4384] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.044 [INFO][4384] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.056 [INFO][4384] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.068 [INFO][4384] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.135/26] block=192.168.8.128/26 handle="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.068 [INFO][4384] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.135/26] handle="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.068 [INFO][4384] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:22.172191 containerd[1481]: 2026-04-28 00:16:22.068 [INFO][4384] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.135/26] IPv6=[] ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" HandleID="k8s-pod-network.2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" Apr 28 00:16:22.172770 containerd[1481]: 2026-04-28 00:16:22.076 [INFO][4314] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0", GenerateName:"whisker-5d8cd77c89-", Namespace:"calico-system", SelfLink:"", UID:"bd6025bd-4014-436d-aeea-02adcdff7a99", ResourceVersion:"945", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5d8cd77c89", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"whisker-5d8cd77c89-ccd2x", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.8.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali28ee2b32562", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:22.172770 containerd[1481]: 2026-04-28 00:16:22.076 [INFO][4314] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.135/32] ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" Apr 28 00:16:22.172770 containerd[1481]: 2026-04-28 00:16:22.076 [INFO][4314] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali28ee2b32562 ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" Apr 28 00:16:22.172770 containerd[1481]: 2026-04-28 00:16:22.111 [INFO][4314] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" Apr 28 00:16:22.172770 containerd[1481]: 2026-04-28 00:16:22.120 [INFO][4314] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0", GenerateName:"whisker-5d8cd77c89-", Namespace:"calico-system", SelfLink:"", UID:"bd6025bd-4014-436d-aeea-02adcdff7a99", ResourceVersion:"945", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5d8cd77c89", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb", Pod:"whisker-5d8cd77c89-ccd2x", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.8.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali28ee2b32562", MAC:"96:51:05:81:50:6c", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:22.172770 containerd[1481]: 2026-04-28 00:16:22.165 [INFO][4314] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb" Namespace="calico-system" Pod="whisker-5d8cd77c89-ccd2x" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--5d8cd77c89--ccd2x-eth0" Apr 28 00:16:22.183950 kubelet[2581]: I0428 00:16:22.183142 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-mqzhw" podStartSLOduration=32.183122147 podStartE2EDuration="32.183122147s" podCreationTimestamp="2026-04-28 00:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-28 00:16:22.165648548 +0000 UTC m=+36.511742739" watchObservedRunningTime="2026-04-28 00:16:22.183122147 +0000 UTC m=+36.529216378" Apr 28 00:16:22.215764 containerd[1481]: time="2026-04-28T00:16:22.215452134Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:22.215764 containerd[1481]: time="2026-04-28T00:16:22.215522095Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:22.215764 containerd[1481]: time="2026-04-28T00:16:22.215542215Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:22.215764 containerd[1481]: time="2026-04-28T00:16:22.215650615Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:22.269070 systemd[1]: Started cri-containerd-2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb.scope - libcontainer container 2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb. Apr 28 00:16:22.280334 containerd[1481]: time="2026-04-28T00:16:22.279064904Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-dcppv,Uid:7335aad5-4b0b-4297-8eed-20f9df1b81e5,Namespace:calico-system,Attempt:1,} returns sandbox id \"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05\"" Apr 28 00:16:22.311253 containerd[1481]: time="2026-04-28T00:16:22.311212450Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-57885fdd4c-glmtw,Uid:c8d091df-86b9-498f-8893-12fb75b0c30e,Namespace:calico-system,Attempt:1,} returns sandbox id \"679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b\"" Apr 28 00:16:22.313913 containerd[1481]: time="2026-04-28T00:16:22.313567861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-8f7677868-s4r9s,Uid:d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b,Namespace:calico-system,Attempt:1,} returns sandbox id \"480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16\"" Apr 28 00:16:22.393329 containerd[1481]: time="2026-04-28T00:16:22.392911942Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d8cd77c89-ccd2x,Uid:bd6025bd-4014-436d-aeea-02adcdff7a99,Namespace:calico-system,Attempt:0,} returns sandbox id \"2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb\"" Apr 28 00:16:22.483913 systemd-networkd[1371]: calid189784c50e: Gained IPv6LL Apr 28 00:16:22.611854 systemd-networkd[1371]: cali7413111db7d: Gained IPv6LL Apr 28 00:16:22.678810 kernel: calico-node[4110]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Apr 28 00:16:22.932186 systemd-networkd[1371]: cali0eb5e55d91c: Gained IPv6LL Apr 28 00:16:23.126458 systemd-networkd[1371]: vxlan.calico: Link UP Apr 28 00:16:23.126469 systemd-networkd[1371]: vxlan.calico: Gained carrier Apr 28 00:16:23.188905 systemd-networkd[1371]: cali0e4558a8691: Gained IPv6LL Apr 28 00:16:23.699910 systemd-networkd[1371]: cali2d5552d3884: Gained IPv6LL Apr 28 00:16:23.700443 systemd-networkd[1371]: cali28ee2b32562: Gained IPv6LL Apr 28 00:16:23.764036 systemd-networkd[1371]: calib31e86a8995: Gained IPv6LL Apr 28 00:16:24.470269 systemd-networkd[1371]: vxlan.calico: Gained IPv6LL Apr 28 00:16:24.947221 containerd[1481]: time="2026-04-28T00:16:24.947052521Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:24.949027 containerd[1481]: time="2026-04-28T00:16:24.948651848Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.5: active requests=0, bytes read=42617669" Apr 28 00:16:24.950737 containerd[1481]: time="2026-04-28T00:16:24.950302534Z" level=info msg="ImageCreate event name:\"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:24.954765 containerd[1481]: time="2026-04-28T00:16:24.954235310Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:78a11eeba8e8a02ecd6014bc8260180819ee7005f9eacb364b9595d1e4b166e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:24.956515 containerd[1481]: time="2026-04-28T00:16:24.956456279Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" with image id \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:78a11eeba8e8a02ecd6014bc8260180819ee7005f9eacb364b9595d1e4b166e1\", size \"45193324\" in 3.45161305s" Apr 28 00:16:24.956515 containerd[1481]: time="2026-04-28T00:16:24.956507879Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" returns image reference \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\"" Apr 28 00:16:24.960989 containerd[1481]: time="2026-04-28T00:16:24.960764256Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\"" Apr 28 00:16:24.964443 containerd[1481]: time="2026-04-28T00:16:24.964261670Z" level=info msg="CreateContainer within sandbox \"1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 28 00:16:24.986997 containerd[1481]: time="2026-04-28T00:16:24.986948121Z" level=info msg="CreateContainer within sandbox \"1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"70d7c91f109d006bf77be537abe03b419b071258cf2ef1a4569152d0ee71f55a\"" Apr 28 00:16:24.988706 containerd[1481]: time="2026-04-28T00:16:24.988198326Z" level=info msg="StartContainer for \"70d7c91f109d006bf77be537abe03b419b071258cf2ef1a4569152d0ee71f55a\"" Apr 28 00:16:25.024918 systemd[1]: Started cri-containerd-70d7c91f109d006bf77be537abe03b419b071258cf2ef1a4569152d0ee71f55a.scope - libcontainer container 70d7c91f109d006bf77be537abe03b419b071258cf2ef1a4569152d0ee71f55a. Apr 28 00:16:25.073702 containerd[1481]: time="2026-04-28T00:16:25.073409048Z" level=info msg="StartContainer for \"70d7c91f109d006bf77be537abe03b419b071258cf2ef1a4569152d0ee71f55a\" returns successfully" Apr 28 00:16:25.166562 kubelet[2581]: I0428 00:16:25.166373 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-674b8b8f7c-xw46n" podStartSLOduration=18.702427971 podStartE2EDuration="22.166354037s" podCreationTimestamp="2026-04-28 00:16:03 +0000 UTC" firstStartedPulling="2026-04-28 00:16:21.496067107 +0000 UTC m=+35.842161298" lastFinishedPulling="2026-04-28 00:16:24.959993173 +0000 UTC m=+39.306087364" observedRunningTime="2026-04-28 00:16:25.165841475 +0000 UTC m=+39.511935666" watchObservedRunningTime="2026-04-28 00:16:25.166354037 +0000 UTC m=+39.512448188" Apr 28 00:16:25.352412 containerd[1481]: time="2026-04-28T00:16:25.352324374Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:25.354986 containerd[1481]: time="2026-04-28T00:16:25.354125141Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.5: active requests=0, bytes read=77" Apr 28 00:16:25.357211 containerd[1481]: time="2026-04-28T00:16:25.357173393Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" with image id \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:78a11eeba8e8a02ecd6014bc8260180819ee7005f9eacb364b9595d1e4b166e1\", size \"45193324\" in 396.367977ms" Apr 28 00:16:25.357211 containerd[1481]: time="2026-04-28T00:16:25.357213113Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.5\" returns image reference \"sha256:3c1e1bbd22dcb1019213c98ef14b99d423455fa7cf8c3a9791619bc5605ccefd\"" Apr 28 00:16:25.358834 containerd[1481]: time="2026-04-28T00:16:25.358802119Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.5\"" Apr 28 00:16:25.362891 containerd[1481]: time="2026-04-28T00:16:25.362858854Z" level=info msg="CreateContainer within sandbox \"7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 28 00:16:25.390000 containerd[1481]: time="2026-04-28T00:16:25.389948796Z" level=info msg="CreateContainer within sandbox \"7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"9023af0cbedd618e2b97a1b9b5fd9ded0ddbb2565f3d13833cb5cc837e293294\"" Apr 28 00:16:25.393058 containerd[1481]: time="2026-04-28T00:16:25.392917007Z" level=info msg="StartContainer for \"9023af0cbedd618e2b97a1b9b5fd9ded0ddbb2565f3d13833cb5cc837e293294\"" Apr 28 00:16:25.422908 systemd[1]: Started cri-containerd-9023af0cbedd618e2b97a1b9b5fd9ded0ddbb2565f3d13833cb5cc837e293294.scope - libcontainer container 9023af0cbedd618e2b97a1b9b5fd9ded0ddbb2565f3d13833cb5cc837e293294. Apr 28 00:16:25.459379 containerd[1481]: time="2026-04-28T00:16:25.459330976Z" level=info msg="StartContainer for \"9023af0cbedd618e2b97a1b9b5fd9ded0ddbb2565f3d13833cb5cc837e293294\" returns successfully" Apr 28 00:16:26.155139 kubelet[2581]: I0428 00:16:26.154696 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 28 00:16:26.982255 containerd[1481]: time="2026-04-28T00:16:26.981957576Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:26.984840 containerd[1481]: time="2026-04-28T00:16:26.983801023Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.5: active requests=0, bytes read=7895994" Apr 28 00:16:26.985239 containerd[1481]: time="2026-04-28T00:16:26.985177108Z" level=info msg="ImageCreate event name:\"sha256:c84299759d8605dff0cc2ebb16a8c098e7266501883bb302cd068ecf668128a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:26.991346 containerd[1481]: time="2026-04-28T00:16:26.988944281Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e8a5b44388a309910946072582b1a1f283c52cf73e9825179235d934447c8b7d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:26.991346 containerd[1481]: time="2026-04-28T00:16:26.989777364Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.5\" with image id \"sha256:c84299759d8605dff0cc2ebb16a8c098e7266501883bb302cd068ecf668128a6\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e8a5b44388a309910946072582b1a1f283c52cf73e9825179235d934447c8b7d\", size \"10471633\" in 1.630938805s" Apr 28 00:16:26.991346 containerd[1481]: time="2026-04-28T00:16:26.989810884Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.5\" returns image reference \"sha256:c84299759d8605dff0cc2ebb16a8c098e7266501883bb302cd068ecf668128a6\"" Apr 28 00:16:26.993366 containerd[1481]: time="2026-04-28T00:16:26.993242896Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.5\"" Apr 28 00:16:26.998376 containerd[1481]: time="2026-04-28T00:16:26.997660031Z" level=info msg="CreateContainer within sandbox \"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Apr 28 00:16:27.023336 containerd[1481]: time="2026-04-28T00:16:27.023070756Z" level=info msg="CreateContainer within sandbox \"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"ba21e25cbb2070a495ed3f0a6b88c12f9c928157c1fbce4cb268223c38eb6b78\"" Apr 28 00:16:27.024174 containerd[1481]: time="2026-04-28T00:16:27.023923919Z" level=info msg="StartContainer for \"ba21e25cbb2070a495ed3f0a6b88c12f9c928157c1fbce4cb268223c38eb6b78\"" Apr 28 00:16:27.074899 systemd[1]: Started cri-containerd-ba21e25cbb2070a495ed3f0a6b88c12f9c928157c1fbce4cb268223c38eb6b78.scope - libcontainer container ba21e25cbb2070a495ed3f0a6b88c12f9c928157c1fbce4cb268223c38eb6b78. Apr 28 00:16:27.107432 containerd[1481]: time="2026-04-28T00:16:27.107378874Z" level=info msg="StartContainer for \"ba21e25cbb2070a495ed3f0a6b88c12f9c928157c1fbce4cb268223c38eb6b78\" returns successfully" Apr 28 00:16:27.159504 kubelet[2581]: I0428 00:16:27.159452 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 28 00:16:29.145417 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3162097642.mount: Deactivated successfully. Apr 28 00:16:29.503593 containerd[1481]: time="2026-04-28T00:16:29.503530085Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:29.506096 containerd[1481]: time="2026-04-28T00:16:29.505223610Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.5: active requests=0, bytes read=48513326" Apr 28 00:16:29.516938 containerd[1481]: time="2026-04-28T00:16:29.516866844Z" level=info msg="ImageCreate event name:\"sha256:f556d75d96fa1483cf593e71a7d71a551e78433f43c12badd65e95187cd0fced\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:29.520435 containerd[1481]: time="2026-04-28T00:16:29.520367774Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:edfd1b6c377013f23afd5e76cb975b6cb59d1bc6554f79c0719d617f8dd0468e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:29.521417 containerd[1481]: time="2026-04-28T00:16:29.521371417Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.5\" with image id \"sha256:f556d75d96fa1483cf593e71a7d71a551e78433f43c12badd65e95187cd0fced\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:edfd1b6c377013f23afd5e76cb975b6cb59d1bc6554f79c0719d617f8dd0468e\", size \"48513172\" in 2.528086561s" Apr 28 00:16:29.521417 containerd[1481]: time="2026-04-28T00:16:29.521415737Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.5\" returns image reference \"sha256:f556d75d96fa1483cf593e71a7d71a551e78433f43c12badd65e95187cd0fced\"" Apr 28 00:16:29.529610 containerd[1481]: time="2026-04-28T00:16:29.529368760Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\"" Apr 28 00:16:29.540791 containerd[1481]: time="2026-04-28T00:16:29.540745033Z" level=info msg="CreateContainer within sandbox \"679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Apr 28 00:16:29.559473 containerd[1481]: time="2026-04-28T00:16:29.559346087Z" level=info msg="CreateContainer within sandbox \"679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0\"" Apr 28 00:16:29.564727 containerd[1481]: time="2026-04-28T00:16:29.564049500Z" level=info msg="StartContainer for \"1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0\"" Apr 28 00:16:29.604216 systemd[1]: Started cri-containerd-1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0.scope - libcontainer container 1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0. Apr 28 00:16:29.641967 containerd[1481]: time="2026-04-28T00:16:29.641582365Z" level=info msg="StartContainer for \"1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0\" returns successfully" Apr 28 00:16:30.207701 kubelet[2581]: I0428 00:16:30.207205 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-674b8b8f7c-jhxz2" podStartSLOduration=23.993970583 podStartE2EDuration="27.207185646s" podCreationTimestamp="2026-04-28 00:16:03 +0000 UTC" firstStartedPulling="2026-04-28 00:16:22.144829253 +0000 UTC m=+36.490923404" lastFinishedPulling="2026-04-28 00:16:25.358044316 +0000 UTC m=+39.704138467" observedRunningTime="2026-04-28 00:16:26.173397293 +0000 UTC m=+40.519491484" watchObservedRunningTime="2026-04-28 00:16:30.207185646 +0000 UTC m=+44.553279837" Apr 28 00:16:31.277226 kubelet[2581]: I0428 00:16:31.276975 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-57885fdd4c-glmtw" podStartSLOduration=20.06479042 podStartE2EDuration="27.276661424s" podCreationTimestamp="2026-04-28 00:16:04 +0000 UTC" firstStartedPulling="2026-04-28 00:16:22.316070992 +0000 UTC m=+36.662165183" lastFinishedPulling="2026-04-28 00:16:29.527942036 +0000 UTC m=+43.874036187" observedRunningTime="2026-04-28 00:16:30.209660773 +0000 UTC m=+44.555754964" watchObservedRunningTime="2026-04-28 00:16:31.276661424 +0000 UTC m=+45.622755615" Apr 28 00:16:31.899108 containerd[1481]: time="2026-04-28T00:16:31.899055129Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:31.900863 containerd[1481]: time="2026-04-28T00:16:31.900794893Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.5: active requests=0, bytes read=46169343" Apr 28 00:16:31.901936 containerd[1481]: time="2026-04-28T00:16:31.901818416Z" level=info msg="ImageCreate event name:\"sha256:f3ba40f705afacb15a8a2f5b02c08a912321f045220eb8f8f1f5ca51f129741a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:31.907032 containerd[1481]: time="2026-04-28T00:16:31.906507948Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5fa7fb7e707d54479cd5d93cfe42352076b805f36560df457b53701d9e738d72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:31.907955 containerd[1481]: time="2026-04-28T00:16:31.907907511Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\" with image id \"sha256:f3ba40f705afacb15a8a2f5b02c08a912321f045220eb8f8f1f5ca51f129741a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5fa7fb7e707d54479cd5d93cfe42352076b805f36560df457b53701d9e738d72\", size \"48744950\" in 2.378494631s" Apr 28 00:16:31.907955 containerd[1481]: time="2026-04-28T00:16:31.907949352Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.5\" returns image reference \"sha256:f3ba40f705afacb15a8a2f5b02c08a912321f045220eb8f8f1f5ca51f129741a\"" Apr 28 00:16:31.910285 containerd[1481]: time="2026-04-28T00:16:31.910234957Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.5\"" Apr 28 00:16:31.942917 containerd[1481]: time="2026-04-28T00:16:31.942720160Z" level=info msg="CreateContainer within sandbox \"480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Apr 28 00:16:31.962712 containerd[1481]: time="2026-04-28T00:16:31.962586611Z" level=info msg="CreateContainer within sandbox \"480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"1ab6e19bb41686e17c19a571e61ee9509afc903ffbe4998804f4a5ca80ff6e16\"" Apr 28 00:16:31.965354 containerd[1481]: time="2026-04-28T00:16:31.965306898Z" level=info msg="StartContainer for \"1ab6e19bb41686e17c19a571e61ee9509afc903ffbe4998804f4a5ca80ff6e16\"" Apr 28 00:16:32.048984 systemd[1]: Started cri-containerd-1ab6e19bb41686e17c19a571e61ee9509afc903ffbe4998804f4a5ca80ff6e16.scope - libcontainer container 1ab6e19bb41686e17c19a571e61ee9509afc903ffbe4998804f4a5ca80ff6e16. Apr 28 00:16:32.091625 containerd[1481]: time="2026-04-28T00:16:32.090571842Z" level=info msg="StartContainer for \"1ab6e19bb41686e17c19a571e61ee9509afc903ffbe4998804f4a5ca80ff6e16\" returns successfully" Apr 28 00:16:32.227366 kubelet[2581]: I0428 00:16:32.227299 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-8f7677868-s4r9s" podStartSLOduration=17.634088567 podStartE2EDuration="27.227278649s" podCreationTimestamp="2026-04-28 00:16:05 +0000 UTC" firstStartedPulling="2026-04-28 00:16:22.316160153 +0000 UTC m=+36.662254344" lastFinishedPulling="2026-04-28 00:16:31.909350235 +0000 UTC m=+46.255444426" observedRunningTime="2026-04-28 00:16:32.222710438 +0000 UTC m=+46.568804629" watchObservedRunningTime="2026-04-28 00:16:32.227278649 +0000 UTC m=+46.573372840" Apr 28 00:16:32.779355 containerd[1481]: time="2026-04-28T00:16:32.779293646Z" level=info msg="StopPodSandbox for \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\"" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.846 [INFO][5024] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.847 [INFO][5024] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" iface="eth0" netns="/var/run/netns/cni-b9a4b397-dda1-d116-affe-1e5041bba926" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.848 [INFO][5024] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" iface="eth0" netns="/var/run/netns/cni-b9a4b397-dda1-d116-affe-1e5041bba926" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.848 [INFO][5024] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" iface="eth0" netns="/var/run/netns/cni-b9a4b397-dda1-d116-affe-1e5041bba926" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.848 [INFO][5024] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.848 [INFO][5024] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.873 [INFO][5031] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.873 [INFO][5031] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.874 [INFO][5031] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.884 [WARNING][5031] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.884 [INFO][5031] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.887 [INFO][5031] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:32.892621 containerd[1481]: 2026-04-28 00:16:32.889 [INFO][5024] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:32.893286 containerd[1481]: time="2026-04-28T00:16:32.892843597Z" level=info msg="TearDown network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\" successfully" Apr 28 00:16:32.893286 containerd[1481]: time="2026-04-28T00:16:32.892875117Z" level=info msg="StopPodSandbox for \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\" returns successfully" Apr 28 00:16:32.895144 containerd[1481]: time="2026-04-28T00:16:32.894863522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-zws2j,Uid:a74e08b1-013f-45ee-bd93-5b8aae9d758d,Namespace:kube-system,Attempt:1,}" Apr 28 00:16:32.923682 systemd[1]: run-netns-cni\x2db9a4b397\x2ddda1\x2dd116\x2daffe\x2d1e5041bba926.mount: Deactivated successfully. Apr 28 00:16:33.061822 systemd-networkd[1371]: cali288f9bad840: Link UP Apr 28 00:16:33.062096 systemd-networkd[1371]: cali288f9bad840: Gained carrier Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:32.962 [INFO][5038] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0 coredns-674b8bbfcf- kube-system a74e08b1-013f-45ee-bd93-5b8aae9d758d 1044 0 2026-04-28 00:15:50 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081-3-7-n-51c70c830d coredns-674b8bbfcf-zws2j eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali288f9bad840 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:32.962 [INFO][5038] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:32.990 [INFO][5050] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" HandleID="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.003 [INFO][5050] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" HandleID="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002ffc00), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081-3-7-n-51c70c830d", "pod":"coredns-674b8bbfcf-zws2j", "timestamp":"2026-04-28 00:16:32.99044387 +0000 UTC"}, Hostname:"ci-4081-3-7-n-51c70c830d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40002ecf20)} Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.003 [INFO][5050] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.003 [INFO][5050] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.003 [INFO][5050] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081-3-7-n-51c70c830d' Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.007 [INFO][5050] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.014 [INFO][5050] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.021 [INFO][5050] ipam/ipam.go 526: Trying affinity for 192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.025 [INFO][5050] ipam/ipam.go 160: Attempting to load block cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.029 [INFO][5050] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.8.128/26 host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.029 [INFO][5050] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.8.128/26 handle="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.032 [INFO][5050] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052 Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.039 [INFO][5050] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.8.128/26 handle="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.048 [INFO][5050] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.8.136/26] block=192.168.8.128/26 handle="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.049 [INFO][5050] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.8.136/26] handle="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" host="ci-4081-3-7-n-51c70c830d" Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.049 [INFO][5050] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:33.083554 containerd[1481]: 2026-04-28 00:16:33.049 [INFO][5050] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.8.136/26] IPv6=[] ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" HandleID="k8s-pod-network.2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:33.086195 containerd[1481]: 2026-04-28 00:16:33.053 [INFO][5038] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a74e08b1-013f-45ee-bd93-5b8aae9d758d", ResourceVersion:"1044", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"", Pod:"coredns-674b8bbfcf-zws2j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali288f9bad840", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:33.086195 containerd[1481]: 2026-04-28 00:16:33.053 [INFO][5038] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.8.136/32] ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:33.086195 containerd[1481]: 2026-04-28 00:16:33.053 [INFO][5038] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali288f9bad840 ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:33.086195 containerd[1481]: 2026-04-28 00:16:33.063 [INFO][5038] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:33.086195 containerd[1481]: 2026-04-28 00:16:33.063 [INFO][5038] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a74e08b1-013f-45ee-bd93-5b8aae9d758d", ResourceVersion:"1044", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052", Pod:"coredns-674b8bbfcf-zws2j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali288f9bad840", MAC:"16:03:ef:d6:ae:85", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:33.086195 containerd[1481]: 2026-04-28 00:16:33.078 [INFO][5038] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052" Namespace="kube-system" Pod="coredns-674b8bbfcf-zws2j" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:33.120851 containerd[1481]: time="2026-04-28T00:16:33.116760955Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 28 00:16:33.120851 containerd[1481]: time="2026-04-28T00:16:33.116824995Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 28 00:16:33.120851 containerd[1481]: time="2026-04-28T00:16:33.116836475Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:33.120851 containerd[1481]: time="2026-04-28T00:16:33.116921915Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 28 00:16:33.147916 systemd[1]: Started cri-containerd-2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052.scope - libcontainer container 2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052. Apr 28 00:16:33.188171 containerd[1481]: time="2026-04-28T00:16:33.188124074Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-zws2j,Uid:a74e08b1-013f-45ee-bd93-5b8aae9d758d,Namespace:kube-system,Attempt:1,} returns sandbox id \"2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052\"" Apr 28 00:16:33.206453 containerd[1481]: time="2026-04-28T00:16:33.205812194Z" level=info msg="CreateContainer within sandbox \"2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 28 00:16:33.286761 containerd[1481]: time="2026-04-28T00:16:33.286664055Z" level=info msg="CreateContainer within sandbox \"2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"e685ea93f2de9bb3f80466e8d4eb8f082004d12813548f9f39c3270ea8d7632a\"" Apr 28 00:16:33.288565 containerd[1481]: time="2026-04-28T00:16:33.288523939Z" level=info msg="StartContainer for \"e685ea93f2de9bb3f80466e8d4eb8f082004d12813548f9f39c3270ea8d7632a\"" Apr 28 00:16:33.335647 systemd[1]: Started cri-containerd-e685ea93f2de9bb3f80466e8d4eb8f082004d12813548f9f39c3270ea8d7632a.scope - libcontainer container e685ea93f2de9bb3f80466e8d4eb8f082004d12813548f9f39c3270ea8d7632a. Apr 28 00:16:33.377079 containerd[1481]: time="2026-04-28T00:16:33.376819017Z" level=info msg="StartContainer for \"e685ea93f2de9bb3f80466e8d4eb8f082004d12813548f9f39c3270ea8d7632a\" returns successfully" Apr 28 00:16:33.535892 containerd[1481]: time="2026-04-28T00:16:33.535845012Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:33.537980 containerd[1481]: time="2026-04-28T00:16:33.537941457Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.5: active requests=0, bytes read=5896864" Apr 28 00:16:33.540311 containerd[1481]: time="2026-04-28T00:16:33.539160660Z" level=info msg="ImageCreate event name:\"sha256:a47d4844a7d3a4350ed0ac1bc7a5e68be5c0d8a9b81906debd805ec9c4deec82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:33.542224 containerd[1481]: time="2026-04-28T00:16:33.542195587Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:b143cf26c347546feabb95cec04a2349f5ae297830cc54fdc2578b89d1a3e021\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:33.543873 containerd[1481]: time="2026-04-28T00:16:33.543827030Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.5\" with image id \"sha256:a47d4844a7d3a4350ed0ac1bc7a5e68be5c0d8a9b81906debd805ec9c4deec82\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:b143cf26c347546feabb95cec04a2349f5ae297830cc54fdc2578b89d1a3e021\", size \"8472495\" in 1.633547113s" Apr 28 00:16:33.544013 containerd[1481]: time="2026-04-28T00:16:33.543995391Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.5\" returns image reference \"sha256:a47d4844a7d3a4350ed0ac1bc7a5e68be5c0d8a9b81906debd805ec9c4deec82\"" Apr 28 00:16:33.547254 containerd[1481]: time="2026-04-28T00:16:33.547220758Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\"" Apr 28 00:16:33.551828 containerd[1481]: time="2026-04-28T00:16:33.551772288Z" level=info msg="CreateContainer within sandbox \"2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Apr 28 00:16:33.577404 containerd[1481]: time="2026-04-28T00:16:33.577354585Z" level=info msg="CreateContainer within sandbox \"2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"de1bb0161b9dba9a15584bfda67fa2e1213ada29b133ccac382c4cef0b31df14\"" Apr 28 00:16:33.578843 containerd[1481]: time="2026-04-28T00:16:33.578809349Z" level=info msg="StartContainer for \"de1bb0161b9dba9a15584bfda67fa2e1213ada29b133ccac382c4cef0b31df14\"" Apr 28 00:16:33.612884 systemd[1]: Started cri-containerd-de1bb0161b9dba9a15584bfda67fa2e1213ada29b133ccac382c4cef0b31df14.scope - libcontainer container de1bb0161b9dba9a15584bfda67fa2e1213ada29b133ccac382c4cef0b31df14. Apr 28 00:16:33.665891 containerd[1481]: time="2026-04-28T00:16:33.665842023Z" level=info msg="StartContainer for \"de1bb0161b9dba9a15584bfda67fa2e1213ada29b133ccac382c4cef0b31df14\" returns successfully" Apr 28 00:16:33.925096 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3021331304.mount: Deactivated successfully. Apr 28 00:16:34.252633 kubelet[2581]: I0428 00:16:34.252542 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-zws2j" podStartSLOduration=44.252522421 podStartE2EDuration="44.252522421s" podCreationTimestamp="2026-04-28 00:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-28 00:16:34.23786611 +0000 UTC m=+48.583960301" watchObservedRunningTime="2026-04-28 00:16:34.252522421 +0000 UTC m=+48.598616612" Apr 28 00:16:35.028031 systemd-networkd[1371]: cali288f9bad840: Gained IPv6LL Apr 28 00:16:35.174991 containerd[1481]: time="2026-04-28T00:16:35.174926413Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:35.176749 containerd[1481]: time="2026-04-28T00:16:35.176660657Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5: active requests=0, bytes read=12456618" Apr 28 00:16:35.177730 containerd[1481]: time="2026-04-28T00:16:35.177451378Z" level=info msg="ImageCreate event name:\"sha256:a127885d176e495b4edc6e0c0309c6570e4d776444937bfdc565fac5a13d8b3f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:35.180581 containerd[1481]: time="2026-04-28T00:16:35.180370504Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:26849483b0c4d797a8ff818d988924bdf696996ca559c8c56b647aaaf70a448a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:35.181484 containerd[1481]: time="2026-04-28T00:16:35.181356586Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\" with image id \"sha256:a127885d176e495b4edc6e0c0309c6570e4d776444937bfdc565fac5a13d8b3f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:26849483b0c4d797a8ff818d988924bdf696996ca559c8c56b647aaaf70a448a\", size \"15032209\" in 1.633482147s" Apr 28 00:16:35.181484 containerd[1481]: time="2026-04-28T00:16:35.181396626Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.5\" returns image reference \"sha256:a127885d176e495b4edc6e0c0309c6570e4d776444937bfdc565fac5a13d8b3f\"" Apr 28 00:16:35.184559 containerd[1481]: time="2026-04-28T00:16:35.184083511Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\"" Apr 28 00:16:35.190530 containerd[1481]: time="2026-04-28T00:16:35.190004563Z" level=info msg="CreateContainer within sandbox \"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Apr 28 00:16:35.218417 containerd[1481]: time="2026-04-28T00:16:35.218367459Z" level=info msg="CreateContainer within sandbox \"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"460a8bd6e4681a3a69069d21b227c4129eccafbe2b009f7d68d29e60b9c97fb9\"" Apr 28 00:16:35.222860 containerd[1481]: time="2026-04-28T00:16:35.222006986Z" level=info msg="StartContainer for \"460a8bd6e4681a3a69069d21b227c4129eccafbe2b009f7d68d29e60b9c97fb9\"" Apr 28 00:16:35.267922 systemd[1]: Started cri-containerd-460a8bd6e4681a3a69069d21b227c4129eccafbe2b009f7d68d29e60b9c97fb9.scope - libcontainer container 460a8bd6e4681a3a69069d21b227c4129eccafbe2b009f7d68d29e60b9c97fb9. Apr 28 00:16:35.308652 containerd[1481]: time="2026-04-28T00:16:35.308393716Z" level=info msg="StartContainer for \"460a8bd6e4681a3a69069d21b227c4129eccafbe2b009f7d68d29e60b9c97fb9\" returns successfully" Apr 28 00:16:35.892164 kubelet[2581]: I0428 00:16:35.892134 2581 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Apr 28 00:16:35.892164 kubelet[2581]: I0428 00:16:35.892169 2581 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Apr 28 00:16:36.259253 kubelet[2581]: I0428 00:16:36.259098 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-dcppv" podStartSLOduration=19.362314182 podStartE2EDuration="32.259077794s" podCreationTimestamp="2026-04-28 00:16:04 +0000 UTC" firstStartedPulling="2026-04-28 00:16:22.286864219 +0000 UTC m=+36.632958370" lastFinishedPulling="2026-04-28 00:16:35.183627791 +0000 UTC m=+49.529721982" observedRunningTime="2026-04-28 00:16:36.255457987 +0000 UTC m=+50.601552258" watchObservedRunningTime="2026-04-28 00:16:36.259077794 +0000 UTC m=+50.605171985" Apr 28 00:16:36.583693 kubelet[2581]: I0428 00:16:36.583063 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 28 00:16:37.004537 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4100802316.mount: Deactivated successfully. Apr 28 00:16:37.030095 containerd[1481]: time="2026-04-28T00:16:37.030050772Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:37.031382 containerd[1481]: time="2026-04-28T00:16:37.031326855Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.5: active requests=0, bytes read=15624823" Apr 28 00:16:37.032488 containerd[1481]: time="2026-04-28T00:16:37.032375296Z" level=info msg="ImageCreate event name:\"sha256:b6ad9a1ad05ff3a8548f5adf860703add7bc41ef2f24f47e461f1914f73f7c8f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:37.036766 containerd[1481]: time="2026-04-28T00:16:37.036330983Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:0bec142ebaa70bcdda5553c7316abcef9cb60a35c2e3ed16b75f26313de91eed\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 28 00:16:37.037237 containerd[1481]: time="2026-04-28T00:16:37.037204985Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\" with image id \"sha256:b6ad9a1ad05ff3a8548f5adf860703add7bc41ef2f24f47e461f1914f73f7c8f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:0bec142ebaa70bcdda5553c7316abcef9cb60a35c2e3ed16b75f26313de91eed\", size \"15624653\" in 1.853057033s" Apr 28 00:16:37.037295 containerd[1481]: time="2026-04-28T00:16:37.037240105Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.5\" returns image reference \"sha256:b6ad9a1ad05ff3a8548f5adf860703add7bc41ef2f24f47e461f1914f73f7c8f\"" Apr 28 00:16:37.042923 containerd[1481]: time="2026-04-28T00:16:37.042846675Z" level=info msg="CreateContainer within sandbox \"2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Apr 28 00:16:37.056905 containerd[1481]: time="2026-04-28T00:16:37.056857099Z" level=info msg="CreateContainer within sandbox \"2f75145ff8d2b1ca102a90f13e4878e4f9d3cfdf8b0163f37b59230df813ffbb\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"15b3639c8d230c001800bba0b3630b1bef8e08609af159fad081fc05070ea577\"" Apr 28 00:16:37.058713 containerd[1481]: time="2026-04-28T00:16:37.058476142Z" level=info msg="StartContainer for \"15b3639c8d230c001800bba0b3630b1bef8e08609af159fad081fc05070ea577\"" Apr 28 00:16:37.099105 systemd[1]: Started cri-containerd-15b3639c8d230c001800bba0b3630b1bef8e08609af159fad081fc05070ea577.scope - libcontainer container 15b3639c8d230c001800bba0b3630b1bef8e08609af159fad081fc05070ea577. Apr 28 00:16:37.142558 containerd[1481]: time="2026-04-28T00:16:37.142430247Z" level=info msg="StartContainer for \"15b3639c8d230c001800bba0b3630b1bef8e08609af159fad081fc05070ea577\" returns successfully" Apr 28 00:16:37.263882 kubelet[2581]: I0428 00:16:37.263249 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-5d8cd77c89-ccd2x" podStartSLOduration=1.6194707780000002 podStartE2EDuration="16.263192615s" podCreationTimestamp="2026-04-28 00:16:21 +0000 UTC" firstStartedPulling="2026-04-28 00:16:22.39463423 +0000 UTC m=+36.740728421" lastFinishedPulling="2026-04-28 00:16:37.038356067 +0000 UTC m=+51.384450258" observedRunningTime="2026-04-28 00:16:37.25989953 +0000 UTC m=+51.605993721" watchObservedRunningTime="2026-04-28 00:16:37.263192615 +0000 UTC m=+51.609286766" Apr 28 00:16:45.788320 containerd[1481]: time="2026-04-28T00:16:45.788284369Z" level=info msg="StopPodSandbox for \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\"" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.840 [WARNING][5331] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"90d5e942-3524-4477-a485-3644be314c39", ResourceVersion:"979", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c", Pod:"coredns-674b8bbfcf-mqzhw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali0e4558a8691", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.840 [INFO][5331] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.840 [INFO][5331] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" iface="eth0" netns="" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.840 [INFO][5331] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.840 [INFO][5331] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.875 [INFO][5338] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.876 [INFO][5338] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.876 [INFO][5338] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.886 [WARNING][5338] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.886 [INFO][5338] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.888 [INFO][5338] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:45.892613 containerd[1481]: 2026-04-28 00:16:45.890 [INFO][5331] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:45.892613 containerd[1481]: time="2026-04-28T00:16:45.892581676Z" level=info msg="TearDown network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\" successfully" Apr 28 00:16:45.892613 containerd[1481]: time="2026-04-28T00:16:45.892625036Z" level=info msg="StopPodSandbox for \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\" returns successfully" Apr 28 00:16:45.897279 containerd[1481]: time="2026-04-28T00:16:45.897100721Z" level=info msg="RemovePodSandbox for \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\"" Apr 28 00:16:45.903293 containerd[1481]: time="2026-04-28T00:16:45.903109687Z" level=info msg="Forcibly stopping sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\"" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.957 [WARNING][5353] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"90d5e942-3524-4477-a485-3644be314c39", ResourceVersion:"979", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"1458bfea44744a11226051858fe4e7133a3ba803e67343aef80ddd4d6ff3f31c", Pod:"coredns-674b8bbfcf-mqzhw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali0e4558a8691", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.957 [INFO][5353] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.957 [INFO][5353] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" iface="eth0" netns="" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.957 [INFO][5353] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.957 [INFO][5353] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.980 [INFO][5360] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.980 [INFO][5360] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.980 [INFO][5360] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.993 [WARNING][5360] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.994 [INFO][5360] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" HandleID="k8s-pod-network.b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--mqzhw-eth0" Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:45.997 [INFO][5360] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.002884 containerd[1481]: 2026-04-28 00:16:46.000 [INFO][5353] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf" Apr 28 00:16:46.002884 containerd[1481]: time="2026-04-28T00:16:46.002830310Z" level=info msg="TearDown network for sandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\" successfully" Apr 28 00:16:46.018255 containerd[1481]: time="2026-04-28T00:16:46.017994805Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:46.018255 containerd[1481]: time="2026-04-28T00:16:46.018090765Z" level=info msg="RemovePodSandbox \"b16bda5f172f14f53d05f58102b5385481aa538928b855491e75c3ccb9ec1dcf\" returns successfully" Apr 28 00:16:46.019489 containerd[1481]: time="2026-04-28T00:16:46.019448806Z" level=info msg="StopPodSandbox for \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\"" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.076 [WARNING][5374] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0", GenerateName:"calico-kube-controllers-8f7677868-", Namespace:"calico-system", SelfLink:"", UID:"d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b", ResourceVersion:"1051", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"8f7677868", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16", Pod:"calico-kube-controllers-8f7677868-s4r9s", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.8.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2d5552d3884", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.077 [INFO][5374] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.077 [INFO][5374] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" iface="eth0" netns="" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.077 [INFO][5374] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.077 [INFO][5374] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.108 [INFO][5382] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.108 [INFO][5382] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.108 [INFO][5382] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.122 [WARNING][5382] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.122 [INFO][5382] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.125 [INFO][5382] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.130392 containerd[1481]: 2026-04-28 00:16:46.127 [INFO][5374] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.131767 containerd[1481]: time="2026-04-28T00:16:46.130359153Z" level=info msg="TearDown network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\" successfully" Apr 28 00:16:46.132158 containerd[1481]: time="2026-04-28T00:16:46.131753675Z" level=info msg="StopPodSandbox for \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\" returns successfully" Apr 28 00:16:46.132594 containerd[1481]: time="2026-04-28T00:16:46.132560555Z" level=info msg="RemovePodSandbox for \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\"" Apr 28 00:16:46.132809 containerd[1481]: time="2026-04-28T00:16:46.132599795Z" level=info msg="Forcibly stopping sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\"" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.176 [WARNING][5396] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0", GenerateName:"calico-kube-controllers-8f7677868-", Namespace:"calico-system", SelfLink:"", UID:"d1762a6f-fcdd-4c15-ab45-2f2e776b2b7b", ResourceVersion:"1051", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"8f7677868", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"480d389294d5b075f4db49795d50fb0fad8f858e6764ff52097c468f2f5cca16", Pod:"calico-kube-controllers-8f7677868-s4r9s", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.8.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2d5552d3884", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.176 [INFO][5396] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.176 [INFO][5396] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" iface="eth0" netns="" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.176 [INFO][5396] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.176 [INFO][5396] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.200 [INFO][5403] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.200 [INFO][5403] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.200 [INFO][5403] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.213 [WARNING][5403] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.214 [INFO][5403] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" HandleID="k8s-pod-network.a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--kube--controllers--8f7677868--s4r9s-eth0" Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.216 [INFO][5403] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.222960 containerd[1481]: 2026-04-28 00:16:46.219 [INFO][5396] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d" Apr 28 00:16:46.223437 containerd[1481]: time="2026-04-28T00:16:46.223062083Z" level=info msg="TearDown network for sandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\" successfully" Apr 28 00:16:46.235018 containerd[1481]: time="2026-04-28T00:16:46.234966854Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:46.235389 containerd[1481]: time="2026-04-28T00:16:46.235240935Z" level=info msg="RemovePodSandbox \"a34ba63b3c64b03e06b29d0d797759f0050f5af2fee164af4ff96460ac25f87d\" returns successfully" Apr 28 00:16:46.235963 containerd[1481]: time="2026-04-28T00:16:46.235919175Z" level=info msg="StopPodSandbox for \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\"" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.283 [WARNING][5417] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0", GenerateName:"goldmane-57885fdd4c-", Namespace:"calico-system", SelfLink:"", UID:"c8d091df-86b9-498f-8893-12fb75b0c30e", ResourceVersion:"1027", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"57885fdd4c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b", Pod:"goldmane-57885fdd4c-glmtw", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.8.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7413111db7d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.283 [INFO][5417] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.283 [INFO][5417] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" iface="eth0" netns="" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.283 [INFO][5417] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.283 [INFO][5417] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.307 [INFO][5424] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.307 [INFO][5424] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.308 [INFO][5424] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.321 [WARNING][5424] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.321 [INFO][5424] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.323 [INFO][5424] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.327868 containerd[1481]: 2026-04-28 00:16:46.325 [INFO][5417] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.329273 containerd[1481]: time="2026-04-28T00:16:46.327885704Z" level=info msg="TearDown network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\" successfully" Apr 28 00:16:46.329273 containerd[1481]: time="2026-04-28T00:16:46.328085384Z" level=info msg="StopPodSandbox for \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\" returns successfully" Apr 28 00:16:46.329273 containerd[1481]: time="2026-04-28T00:16:46.328770465Z" level=info msg="RemovePodSandbox for \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\"" Apr 28 00:16:46.329273 containerd[1481]: time="2026-04-28T00:16:46.328815745Z" level=info msg="Forcibly stopping sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\"" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.375 [WARNING][5440] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0", GenerateName:"goldmane-57885fdd4c-", Namespace:"calico-system", SelfLink:"", UID:"c8d091df-86b9-498f-8893-12fb75b0c30e", ResourceVersion:"1027", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"57885fdd4c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"679dec3eed6e6f722286f25dc6188419283687ce5ced61d357545b8ea9414d6b", Pod:"goldmane-57885fdd4c-glmtw", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.8.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7413111db7d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.375 [INFO][5440] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.375 [INFO][5440] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" iface="eth0" netns="" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.375 [INFO][5440] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.375 [INFO][5440] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.400 [INFO][5447] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.400 [INFO][5447] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.400 [INFO][5447] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.413 [WARNING][5447] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.413 [INFO][5447] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" HandleID="k8s-pod-network.8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Workload="ci--4081--3--7--n--51c70c830d-k8s-goldmane--57885fdd4c--glmtw-eth0" Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.415 [INFO][5447] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.422501 containerd[1481]: 2026-04-28 00:16:46.418 [INFO][5440] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc" Apr 28 00:16:46.422501 containerd[1481]: time="2026-04-28T00:16:46.422344676Z" level=info msg="TearDown network for sandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\" successfully" Apr 28 00:16:46.428900 containerd[1481]: time="2026-04-28T00:16:46.428489122Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:46.428900 containerd[1481]: time="2026-04-28T00:16:46.428731402Z" level=info msg="RemovePodSandbox \"8fe9f8863e1d5bc876b94797137b9810c4ae787e98a9fd2f1f099a2e3824a4dc\" returns successfully" Apr 28 00:16:46.429541 containerd[1481]: time="2026-04-28T00:16:46.429470643Z" level=info msg="StopPodSandbox for \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\"" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.474 [WARNING][5461] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.474 [INFO][5461] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.474 [INFO][5461] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" iface="eth0" netns="" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.474 [INFO][5461] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.474 [INFO][5461] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.502 [INFO][5468] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.502 [INFO][5468] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.502 [INFO][5468] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.514 [WARNING][5468] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.514 [INFO][5468] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.517 [INFO][5468] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.521140 containerd[1481]: 2026-04-28 00:16:46.519 [INFO][5461] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.522066 containerd[1481]: time="2026-04-28T00:16:46.521234891Z" level=info msg="TearDown network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\" successfully" Apr 28 00:16:46.522066 containerd[1481]: time="2026-04-28T00:16:46.521261011Z" level=info msg="StopPodSandbox for \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\" returns successfully" Apr 28 00:16:46.522762 containerd[1481]: time="2026-04-28T00:16:46.522636653Z" level=info msg="RemovePodSandbox for \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\"" Apr 28 00:16:46.522893 containerd[1481]: time="2026-04-28T00:16:46.522861853Z" level=info msg="Forcibly stopping sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\"" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.572 [WARNING][5482] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" WorkloadEndpoint="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.572 [INFO][5482] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.572 [INFO][5482] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" iface="eth0" netns="" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.572 [INFO][5482] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.572 [INFO][5482] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.598 [INFO][5489] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.598 [INFO][5489] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.598 [INFO][5489] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.609 [WARNING][5489] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.609 [INFO][5489] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" HandleID="k8s-pod-network.1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Workload="ci--4081--3--7--n--51c70c830d-k8s-whisker--7486974bdf--9jnwq-eth0" Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.612 [INFO][5489] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.615795 containerd[1481]: 2026-04-28 00:16:46.613 [INFO][5482] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7" Apr 28 00:16:46.616761 containerd[1481]: time="2026-04-28T00:16:46.615926183Z" level=info msg="TearDown network for sandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\" successfully" Apr 28 00:16:46.621400 containerd[1481]: time="2026-04-28T00:16:46.621205388Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:46.621400 containerd[1481]: time="2026-04-28T00:16:46.621290268Z" level=info msg="RemovePodSandbox \"1b2e0d2cfa93eebf7769ebcf82e88805dcfca4b52628ddc113f3a44fb22648a7\" returns successfully" Apr 28 00:16:46.621823 containerd[1481]: time="2026-04-28T00:16:46.621793629Z" level=info msg="StopPodSandbox for \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\"" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.675 [WARNING][5503] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a74e08b1-013f-45ee-bd93-5b8aae9d758d", ResourceVersion:"1066", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052", Pod:"coredns-674b8bbfcf-zws2j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali288f9bad840", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.676 [INFO][5503] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.676 [INFO][5503] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" iface="eth0" netns="" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.676 [INFO][5503] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.676 [INFO][5503] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.697 [INFO][5510] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.698 [INFO][5510] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.698 [INFO][5510] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.711 [WARNING][5510] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.711 [INFO][5510] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.713 [INFO][5510] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.719240 containerd[1481]: 2026-04-28 00:16:46.716 [INFO][5503] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.719953 containerd[1481]: time="2026-04-28T00:16:46.719294843Z" level=info msg="TearDown network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\" successfully" Apr 28 00:16:46.719953 containerd[1481]: time="2026-04-28T00:16:46.719351323Z" level=info msg="StopPodSandbox for \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\" returns successfully" Apr 28 00:16:46.720172 containerd[1481]: time="2026-04-28T00:16:46.720070164Z" level=info msg="RemovePodSandbox for \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\"" Apr 28 00:16:46.720172 containerd[1481]: time="2026-04-28T00:16:46.720126764Z" level=info msg="Forcibly stopping sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\"" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.762 [WARNING][5524] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a74e08b1-013f-45ee-bd93-5b8aae9d758d", ResourceVersion:"1066", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 15, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"2d72dc79ffd487a9751c3f6d93f4e20a19354f5aa46f7fbe150addb319d47052", Pod:"coredns-674b8bbfcf-zws2j", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.8.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali288f9bad840", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.762 [INFO][5524] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.762 [INFO][5524] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" iface="eth0" netns="" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.762 [INFO][5524] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.762 [INFO][5524] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.783 [INFO][5531] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.783 [INFO][5531] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.783 [INFO][5531] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.793 [WARNING][5531] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.793 [INFO][5531] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" HandleID="k8s-pod-network.cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Workload="ci--4081--3--7--n--51c70c830d-k8s-coredns--674b8bbfcf--zws2j-eth0" Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.794 [INFO][5531] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.800606 containerd[1481]: 2026-04-28 00:16:46.796 [INFO][5524] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4" Apr 28 00:16:46.802981 containerd[1481]: time="2026-04-28T00:16:46.801890123Z" level=info msg="TearDown network for sandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\" successfully" Apr 28 00:16:46.807034 containerd[1481]: time="2026-04-28T00:16:46.806805607Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:46.807034 containerd[1481]: time="2026-04-28T00:16:46.806969088Z" level=info msg="RemovePodSandbox \"cdb4f741f976056bc7b48cb7e06670564e246b349861c877d98c05eaf4f9dfb4\" returns successfully" Apr 28 00:16:46.808804 containerd[1481]: time="2026-04-28T00:16:46.808682649Z" level=info msg="StopPodSandbox for \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\"" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.855 [WARNING][5545] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"f4111884-15d4-472c-b571-63308fbf58dd", ResourceVersion:"993", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e", Pod:"calico-apiserver-674b8b8f7c-xw46n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali0eb5e55d91c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.855 [INFO][5545] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.855 [INFO][5545] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" iface="eth0" netns="" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.855 [INFO][5545] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.855 [INFO][5545] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.880 [INFO][5552] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.880 [INFO][5552] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.880 [INFO][5552] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.891 [WARNING][5552] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.891 [INFO][5552] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.894 [INFO][5552] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:46.899421 containerd[1481]: 2026-04-28 00:16:46.897 [INFO][5545] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:46.899421 containerd[1481]: time="2026-04-28T00:16:46.899255817Z" level=info msg="TearDown network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\" successfully" Apr 28 00:16:46.899421 containerd[1481]: time="2026-04-28T00:16:46.899297577Z" level=info msg="StopPodSandbox for \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\" returns successfully" Apr 28 00:16:46.900441 containerd[1481]: time="2026-04-28T00:16:46.900318018Z" level=info msg="RemovePodSandbox for \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\"" Apr 28 00:16:46.900441 containerd[1481]: time="2026-04-28T00:16:46.900350618Z" level=info msg="Forcibly stopping sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\"" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.961 [WARNING][5567] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"f4111884-15d4-472c-b571-63308fbf58dd", ResourceVersion:"993", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"1f9d4913cdf07f15aae246156a9a963f4dec5ebe7543bc1d9864099a54f0d32e", Pod:"calico-apiserver-674b8b8f7c-xw46n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali0eb5e55d91c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.962 [INFO][5567] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.962 [INFO][5567] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" iface="eth0" netns="" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.962 [INFO][5567] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.962 [INFO][5567] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.986 [INFO][5574] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.987 [INFO][5574] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.987 [INFO][5574] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.999 [WARNING][5574] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:46.999 [INFO][5574] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" HandleID="k8s-pod-network.2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--xw46n-eth0" Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:47.001 [INFO][5574] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:47.006719 containerd[1481]: 2026-04-28 00:16:47.004 [INFO][5567] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e" Apr 28 00:16:47.006719 containerd[1481]: time="2026-04-28T00:16:47.006666960Z" level=info msg="TearDown network for sandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\" successfully" Apr 28 00:16:47.011610 containerd[1481]: time="2026-04-28T00:16:47.011545365Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:47.011773 containerd[1481]: time="2026-04-28T00:16:47.011650845Z" level=info msg="RemovePodSandbox \"2c90bd64d928870a83599f28e7d4ea914ef554cc58d05cda5edc8e3d3cb00a4e\" returns successfully" Apr 28 00:16:47.012311 containerd[1481]: time="2026-04-28T00:16:47.012275045Z" level=info msg="StopPodSandbox for \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\"" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.060 [WARNING][5588] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"9d521bc8-2ea7-4994-aa05-4525d03b6763", ResourceVersion:"1083", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795", Pod:"calico-apiserver-674b8b8f7c-jhxz2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calid189784c50e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.061 [INFO][5588] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.061 [INFO][5588] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" iface="eth0" netns="" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.061 [INFO][5588] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.061 [INFO][5588] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.089 [INFO][5595] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.089 [INFO][5595] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.089 [INFO][5595] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.101 [WARNING][5595] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.101 [INFO][5595] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.104 [INFO][5595] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:47.108563 containerd[1481]: 2026-04-28 00:16:47.106 [INFO][5588] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.109070 containerd[1481]: time="2026-04-28T00:16:47.108605093Z" level=info msg="TearDown network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\" successfully" Apr 28 00:16:47.109070 containerd[1481]: time="2026-04-28T00:16:47.108641933Z" level=info msg="StopPodSandbox for \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\" returns successfully" Apr 28 00:16:47.109587 containerd[1481]: time="2026-04-28T00:16:47.109546854Z" level=info msg="RemovePodSandbox for \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\"" Apr 28 00:16:47.109642 containerd[1481]: time="2026-04-28T00:16:47.109601654Z" level=info msg="Forcibly stopping sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\"" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.152 [WARNING][5609] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0", GenerateName:"calico-apiserver-674b8b8f7c-", Namespace:"calico-system", SelfLink:"", UID:"9d521bc8-2ea7-4994-aa05-4525d03b6763", ResourceVersion:"1083", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"674b8b8f7c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"7f4b8ca3f4186b5111f50f46316df420be9ae20785a868cc05a51524ddbd4795", Pod:"calico-apiserver-674b8b8f7c-jhxz2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.8.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calid189784c50e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.153 [INFO][5609] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.153 [INFO][5609] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" iface="eth0" netns="" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.153 [INFO][5609] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.153 [INFO][5609] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.177 [INFO][5616] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.178 [INFO][5616] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.178 [INFO][5616] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.190 [WARNING][5616] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.190 [INFO][5616] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" HandleID="k8s-pod-network.290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Workload="ci--4081--3--7--n--51c70c830d-k8s-calico--apiserver--674b8b8f7c--jhxz2-eth0" Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.192 [INFO][5616] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:47.196517 containerd[1481]: 2026-04-28 00:16:47.194 [INFO][5609] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950" Apr 28 00:16:47.197035 containerd[1481]: time="2026-04-28T00:16:47.196556293Z" level=info msg="TearDown network for sandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\" successfully" Apr 28 00:16:47.201020 containerd[1481]: time="2026-04-28T00:16:47.200956937Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:47.201398 containerd[1481]: time="2026-04-28T00:16:47.201070057Z" level=info msg="RemovePodSandbox \"290ab4c9bff17096ef8a750d62bb170ebb2669578083c797e6bffa7dde125950\" returns successfully" Apr 28 00:16:47.201613 containerd[1481]: time="2026-04-28T00:16:47.201583617Z" level=info msg="StopPodSandbox for \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\"" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.248 [WARNING][5630] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7335aad5-4b0b-4297-8eed-20f9df1b81e5", ResourceVersion:"1080", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"74865c565", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05", Pod:"csi-node-driver-dcppv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.8.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calib31e86a8995", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.248 [INFO][5630] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.248 [INFO][5630] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" iface="eth0" netns="" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.248 [INFO][5630] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.248 [INFO][5630] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.271 [INFO][5638] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.271 [INFO][5638] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.272 [INFO][5638] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.285 [WARNING][5638] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.285 [INFO][5638] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.288 [INFO][5638] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:47.293322 containerd[1481]: 2026-04-28 00:16:47.291 [INFO][5630] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.293322 containerd[1481]: time="2026-04-28T00:16:47.293244340Z" level=info msg="TearDown network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\" successfully" Apr 28 00:16:47.293322 containerd[1481]: time="2026-04-28T00:16:47.293271940Z" level=info msg="StopPodSandbox for \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\" returns successfully" Apr 28 00:16:47.295242 containerd[1481]: time="2026-04-28T00:16:47.294939942Z" level=info msg="RemovePodSandbox for \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\"" Apr 28 00:16:47.295242 containerd[1481]: time="2026-04-28T00:16:47.294974782Z" level=info msg="Forcibly stopping sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\"" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.342 [WARNING][5652] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7335aad5-4b0b-4297-8eed-20f9df1b81e5", ResourceVersion:"1080", Generation:0, CreationTimestamp:time.Date(2026, time.April, 28, 0, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"74865c565", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081-3-7-n-51c70c830d", ContainerID:"0a1822976a5f8b0e412e3bd6e08a06fba09ab69ecad233cb6e4db53949bc5a05", Pod:"csi-node-driver-dcppv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.8.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calib31e86a8995", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.342 [INFO][5652] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.342 [INFO][5652] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" iface="eth0" netns="" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.342 [INFO][5652] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.342 [INFO][5652] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.369 [INFO][5659] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.369 [INFO][5659] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.370 [INFO][5659] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.383 [WARNING][5659] ipam/ipam_plugin.go 515: Asked to release address but it doesn't exist. Ignoring ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.383 [INFO][5659] ipam/ipam_plugin.go 526: Releasing address using workloadID ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" HandleID="k8s-pod-network.a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Workload="ci--4081--3--7--n--51c70c830d-k8s-csi--node--driver--dcppv-eth0" Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.385 [INFO][5659] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Apr 28 00:16:47.392117 containerd[1481]: 2026-04-28 00:16:47.389 [INFO][5652] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803" Apr 28 00:16:47.392117 containerd[1481]: time="2026-04-28T00:16:47.392095590Z" level=info msg="TearDown network for sandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\" successfully" Apr 28 00:16:47.396601 containerd[1481]: time="2026-04-28T00:16:47.396543914Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 28 00:16:47.397652 containerd[1481]: time="2026-04-28T00:16:47.396625354Z" level=info msg="RemovePodSandbox \"a0ab22a8cff3cc69d2ebc5c4db0cc99f123bdbc161f42e1598a4661daaaed803\" returns successfully" Apr 28 00:16:49.628646 kubelet[2581]: I0428 00:16:49.628135 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 28 00:16:51.121845 systemd[1]: run-containerd-runc-k8s.io-2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049-runc.WoHNCG.mount: Deactivated successfully. Apr 28 00:17:09.357408 systemd[1]: run-containerd-runc-k8s.io-1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0-runc.GbZbUY.mount: Deactivated successfully. Apr 28 00:18:01.213017 systemd[1]: run-containerd-runc-k8s.io-1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0-runc.8Y7DMD.mount: Deactivated successfully. Apr 28 00:18:03.236070 systemd[1]: run-containerd-runc-k8s.io-1ab6e19bb41686e17c19a571e61ee9509afc903ffbe4998804f4a5ca80ff6e16-runc.0kuBcb.mount: Deactivated successfully. Apr 28 00:18:08.789040 systemd[1]: Started sshd@7-178.105.21.241:22-50.85.169.122:39886.service - OpenSSH per-connection server daemon (50.85.169.122:39886). Apr 28 00:18:08.929396 sshd[5961]: Accepted publickey for core from 50.85.169.122 port 39886 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:08.931974 sshd[5961]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:08.938647 systemd-logind[1454]: New session 8 of user core. Apr 28 00:18:08.947005 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 28 00:18:09.157918 sshd[5961]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:09.164325 systemd-logind[1454]: Session 8 logged out. Waiting for processes to exit. Apr 28 00:18:09.166094 systemd[1]: sshd@7-178.105.21.241:22-50.85.169.122:39886.service: Deactivated successfully. Apr 28 00:18:09.169360 systemd[1]: session-8.scope: Deactivated successfully. Apr 28 00:18:09.171011 systemd-logind[1454]: Removed session 8. Apr 28 00:18:09.364380 systemd[1]: run-containerd-runc-k8s.io-1a5d4d39119246823e66b25ab18b04161ce1cd5ff930dc5121e4d48db7ad67e0-runc.qL59rP.mount: Deactivated successfully. Apr 28 00:18:14.196407 systemd[1]: Started sshd@8-178.105.21.241:22-50.85.169.122:55112.service - OpenSSH per-connection server daemon (50.85.169.122:55112). Apr 28 00:18:14.321715 sshd[5995]: Accepted publickey for core from 50.85.169.122 port 55112 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:14.325222 sshd[5995]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:14.330764 systemd-logind[1454]: New session 9 of user core. Apr 28 00:18:14.335939 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 28 00:18:14.520496 sshd[5995]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:14.526601 systemd[1]: sshd@8-178.105.21.241:22-50.85.169.122:55112.service: Deactivated successfully. Apr 28 00:18:14.531049 systemd[1]: session-9.scope: Deactivated successfully. Apr 28 00:18:14.534064 systemd-logind[1454]: Session 9 logged out. Waiting for processes to exit. Apr 28 00:18:14.536291 systemd-logind[1454]: Removed session 9. Apr 28 00:18:19.553983 systemd[1]: Started sshd@9-178.105.21.241:22-50.85.169.122:50188.service - OpenSSH per-connection server daemon (50.85.169.122:50188). Apr 28 00:18:19.673714 sshd[6010]: Accepted publickey for core from 50.85.169.122 port 50188 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:19.677711 sshd[6010]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:19.683966 systemd-logind[1454]: New session 10 of user core. Apr 28 00:18:19.692386 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 28 00:18:19.872807 sshd[6010]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:19.879502 systemd[1]: sshd@9-178.105.21.241:22-50.85.169.122:50188.service: Deactivated successfully. Apr 28 00:18:19.881930 systemd[1]: session-10.scope: Deactivated successfully. Apr 28 00:18:19.883991 systemd-logind[1454]: Session 10 logged out. Waiting for processes to exit. Apr 28 00:18:19.885405 systemd-logind[1454]: Removed session 10. Apr 28 00:18:23.900153 systemd[1]: Started sshd@10-178.105.21.241:22-90.226.146.31:48392.service - OpenSSH per-connection server daemon (90.226.146.31:48392). Apr 28 00:18:24.909132 systemd[1]: Started sshd@11-178.105.21.241:22-50.85.169.122:50198.service - OpenSSH per-connection server daemon (50.85.169.122:50198). Apr 28 00:18:25.039112 sshd[6050]: Accepted publickey for core from 50.85.169.122 port 50198 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:25.042748 sshd[6050]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:25.051053 systemd-logind[1454]: New session 11 of user core. Apr 28 00:18:25.061123 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 28 00:18:25.251059 sshd[6050]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:25.257365 systemd[1]: sshd@11-178.105.21.241:22-50.85.169.122:50198.service: Deactivated successfully. Apr 28 00:18:25.262243 systemd[1]: session-11.scope: Deactivated successfully. Apr 28 00:18:25.264705 systemd-logind[1454]: Session 11 logged out. Waiting for processes to exit. Apr 28 00:18:25.275004 systemd-logind[1454]: Removed session 11. Apr 28 00:18:25.278196 systemd[1]: Started sshd@12-178.105.21.241:22-50.85.169.122:50212.service - OpenSSH per-connection server daemon (50.85.169.122:50212). Apr 28 00:18:25.420772 sshd[6068]: Accepted publickey for core from 50.85.169.122 port 50212 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:25.423658 sshd[6068]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:25.428896 systemd-logind[1454]: New session 12 of user core. Apr 28 00:18:25.432882 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 28 00:18:25.651543 sshd[6068]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:25.655119 systemd[1]: sshd@12-178.105.21.241:22-50.85.169.122:50212.service: Deactivated successfully. Apr 28 00:18:25.658689 systemd[1]: session-12.scope: Deactivated successfully. Apr 28 00:18:25.665444 systemd-logind[1454]: Session 12 logged out. Waiting for processes to exit. Apr 28 00:18:25.682771 systemd[1]: Started sshd@13-178.105.21.241:22-50.85.169.122:50228.service - OpenSSH per-connection server daemon (50.85.169.122:50228). Apr 28 00:18:25.684053 systemd-logind[1454]: Removed session 12. Apr 28 00:18:25.803402 sshd[6080]: Accepted publickey for core from 50.85.169.122 port 50228 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:25.804623 sshd[6080]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:25.812424 systemd-logind[1454]: New session 13 of user core. Apr 28 00:18:25.819898 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 28 00:18:26.006293 sshd[6080]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:26.011945 systemd[1]: sshd@13-178.105.21.241:22-50.85.169.122:50228.service: Deactivated successfully. Apr 28 00:18:26.012209 systemd-logind[1454]: Session 13 logged out. Waiting for processes to exit. Apr 28 00:18:26.015523 systemd[1]: session-13.scope: Deactivated successfully. Apr 28 00:18:26.016988 systemd-logind[1454]: Removed session 13. Apr 28 00:18:30.400608 update_engine[1455]: I20260428 00:18:30.398787 1455 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Apr 28 00:18:30.400608 update_engine[1455]: I20260428 00:18:30.398857 1455 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Apr 28 00:18:30.400608 update_engine[1455]: I20260428 00:18:30.399233 1455 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Apr 28 00:18:30.400608 update_engine[1455]: I20260428 00:18:30.399947 1455 omaha_request_params.cc:62] Current group set to lts Apr 28 00:18:30.400608 update_engine[1455]: I20260428 00:18:30.400298 1455 update_attempter.cc:499] Already updated boot flags. Skipping. Apr 28 00:18:30.400608 update_engine[1455]: I20260428 00:18:30.400326 1455 update_attempter.cc:643] Scheduling an action processor start. Apr 28 00:18:30.400608 update_engine[1455]: I20260428 00:18:30.400354 1455 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 28 00:18:30.402147 update_engine[1455]: I20260428 00:18:30.401830 1455 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Apr 28 00:18:30.402147 update_engine[1455]: I20260428 00:18:30.402004 1455 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 28 00:18:30.402147 update_engine[1455]: I20260428 00:18:30.402024 1455 omaha_request_action.cc:272] Request: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: Apr 28 00:18:30.402147 update_engine[1455]: I20260428 00:18:30.402037 1455 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 28 00:18:30.409012 update_engine[1455]: I20260428 00:18:30.407292 1455 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 28 00:18:30.409012 update_engine[1455]: I20260428 00:18:30.407624 1455 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 28 00:18:30.409372 update_engine[1455]: E20260428 00:18:30.409171 1455 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 28 00:18:30.409372 update_engine[1455]: I20260428 00:18:30.409250 1455 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Apr 28 00:18:30.409481 locksmithd[1488]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Apr 28 00:18:31.040975 systemd[1]: Started sshd@14-178.105.21.241:22-50.85.169.122:52104.service - OpenSSH per-connection server daemon (50.85.169.122:52104). Apr 28 00:18:31.166665 sshd[6092]: Accepted publickey for core from 50.85.169.122 port 52104 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:31.168754 sshd[6092]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:31.173406 systemd-logind[1454]: New session 14 of user core. Apr 28 00:18:31.182018 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 28 00:18:31.360730 sshd[6092]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:31.366769 systemd[1]: sshd@14-178.105.21.241:22-50.85.169.122:52104.service: Deactivated successfully. Apr 28 00:18:31.368977 systemd[1]: session-14.scope: Deactivated successfully. Apr 28 00:18:31.371264 systemd-logind[1454]: Session 14 logged out. Waiting for processes to exit. Apr 28 00:18:31.372376 systemd-logind[1454]: Removed session 14. Apr 28 00:18:33.863899 sshd[6048]: Invalid user admin from 90.226.146.31 port 48392 Apr 28 00:18:34.411127 sshd[6048]: Connection closed by invalid user admin 90.226.146.31 port 48392 [preauth] Apr 28 00:18:34.415258 systemd[1]: sshd@10-178.105.21.241:22-90.226.146.31:48392.service: Deactivated successfully. Apr 28 00:18:36.395211 systemd[1]: Started sshd@15-178.105.21.241:22-50.85.169.122:52118.service - OpenSSH per-connection server daemon (50.85.169.122:52118). Apr 28 00:18:36.514189 sshd[6145]: Accepted publickey for core from 50.85.169.122 port 52118 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:36.516888 sshd[6145]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:36.524505 systemd-logind[1454]: New session 15 of user core. Apr 28 00:18:36.530895 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 28 00:18:36.721828 sshd[6145]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:36.727512 systemd[1]: sshd@15-178.105.21.241:22-50.85.169.122:52118.service: Deactivated successfully. Apr 28 00:18:36.730250 systemd[1]: session-15.scope: Deactivated successfully. Apr 28 00:18:36.731277 systemd-logind[1454]: Session 15 logged out. Waiting for processes to exit. Apr 28 00:18:36.733520 systemd-logind[1454]: Removed session 15. Apr 28 00:18:36.757896 systemd[1]: Started sshd@16-178.105.21.241:22-50.85.169.122:52128.service - OpenSSH per-connection server daemon (50.85.169.122:52128). Apr 28 00:18:36.891829 sshd[6157]: Accepted publickey for core from 50.85.169.122 port 52128 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:36.893785 sshd[6157]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:36.901343 systemd-logind[1454]: New session 16 of user core. Apr 28 00:18:36.910074 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 28 00:18:37.255178 sshd[6157]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:37.259395 systemd[1]: sshd@16-178.105.21.241:22-50.85.169.122:52128.service: Deactivated successfully. Apr 28 00:18:37.264586 systemd[1]: session-16.scope: Deactivated successfully. Apr 28 00:18:37.266265 systemd-logind[1454]: Session 16 logged out. Waiting for processes to exit. Apr 28 00:18:37.267465 systemd-logind[1454]: Removed session 16. Apr 28 00:18:37.285227 systemd[1]: Started sshd@17-178.105.21.241:22-50.85.169.122:52140.service - OpenSSH per-connection server daemon (50.85.169.122:52140). Apr 28 00:18:37.419974 sshd[6168]: Accepted publickey for core from 50.85.169.122 port 52140 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:37.421976 sshd[6168]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:37.429855 systemd-logind[1454]: New session 17 of user core. Apr 28 00:18:37.438063 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 28 00:18:38.301641 sshd[6168]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:38.306450 systemd[1]: sshd@17-178.105.21.241:22-50.85.169.122:52140.service: Deactivated successfully. Apr 28 00:18:38.309527 systemd[1]: session-17.scope: Deactivated successfully. Apr 28 00:18:38.315709 systemd-logind[1454]: Session 17 logged out. Waiting for processes to exit. Apr 28 00:18:38.327366 systemd-logind[1454]: Removed session 17. Apr 28 00:18:38.335030 systemd[1]: Started sshd@18-178.105.21.241:22-50.85.169.122:52156.service - OpenSSH per-connection server daemon (50.85.169.122:52156). Apr 28 00:18:38.467574 sshd[6187]: Accepted publickey for core from 50.85.169.122 port 52156 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:38.469943 sshd[6187]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:38.476406 systemd-logind[1454]: New session 18 of user core. Apr 28 00:18:38.482971 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 28 00:18:38.815494 sshd[6187]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:38.823003 systemd[1]: sshd@18-178.105.21.241:22-50.85.169.122:52156.service: Deactivated successfully. Apr 28 00:18:38.829021 systemd[1]: session-18.scope: Deactivated successfully. Apr 28 00:18:38.842948 systemd-logind[1454]: Session 18 logged out. Waiting for processes to exit. Apr 28 00:18:38.849243 systemd[1]: Started sshd@19-178.105.21.241:22-50.85.169.122:52158.service - OpenSSH per-connection server daemon (50.85.169.122:52158). Apr 28 00:18:38.852208 systemd-logind[1454]: Removed session 18. Apr 28 00:18:38.980500 sshd[6201]: Accepted publickey for core from 50.85.169.122 port 52158 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:38.984203 sshd[6201]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:38.991153 systemd-logind[1454]: New session 19 of user core. Apr 28 00:18:38.996920 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 28 00:18:39.194195 sshd[6201]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:39.200320 systemd-logind[1454]: Session 19 logged out. Waiting for processes to exit. Apr 28 00:18:39.201273 systemd[1]: sshd@19-178.105.21.241:22-50.85.169.122:52158.service: Deactivated successfully. Apr 28 00:18:39.203889 systemd[1]: session-19.scope: Deactivated successfully. Apr 28 00:18:39.205598 systemd-logind[1454]: Removed session 19. Apr 28 00:18:40.398310 update_engine[1455]: I20260428 00:18:40.398162 1455 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 28 00:18:40.398981 update_engine[1455]: I20260428 00:18:40.398542 1455 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 28 00:18:40.398981 update_engine[1455]: I20260428 00:18:40.398933 1455 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 28 00:18:40.399881 update_engine[1455]: E20260428 00:18:40.399812 1455 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 28 00:18:40.399994 update_engine[1455]: I20260428 00:18:40.399907 1455 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Apr 28 00:18:44.228145 systemd[1]: Started sshd@20-178.105.21.241:22-50.85.169.122:56598.service - OpenSSH per-connection server daemon (50.85.169.122:56598). Apr 28 00:18:44.350718 sshd[6215]: Accepted publickey for core from 50.85.169.122 port 56598 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:44.352332 sshd[6215]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:44.359488 systemd-logind[1454]: New session 20 of user core. Apr 28 00:18:44.365960 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 28 00:18:44.539031 sshd[6215]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:44.544487 systemd[1]: sshd@20-178.105.21.241:22-50.85.169.122:56598.service: Deactivated successfully. Apr 28 00:18:44.548903 systemd[1]: session-20.scope: Deactivated successfully. Apr 28 00:18:44.550423 systemd-logind[1454]: Session 20 logged out. Waiting for processes to exit. Apr 28 00:18:44.552502 systemd-logind[1454]: Removed session 20. Apr 28 00:18:49.570117 systemd[1]: Started sshd@21-178.105.21.241:22-50.85.169.122:44984.service - OpenSSH per-connection server daemon (50.85.169.122:44984). Apr 28 00:18:49.698097 sshd[6229]: Accepted publickey for core from 50.85.169.122 port 44984 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:49.700727 sshd[6229]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:49.707364 systemd-logind[1454]: New session 21 of user core. Apr 28 00:18:49.713930 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 28 00:18:49.889246 sshd[6229]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:49.895665 systemd-logind[1454]: Session 21 logged out. Waiting for processes to exit. Apr 28 00:18:49.896379 systemd[1]: sshd@21-178.105.21.241:22-50.85.169.122:44984.service: Deactivated successfully. Apr 28 00:18:49.898579 systemd[1]: session-21.scope: Deactivated successfully. Apr 28 00:18:49.900561 systemd-logind[1454]: Removed session 21. Apr 28 00:18:50.398105 update_engine[1455]: I20260428 00:18:50.397979 1455 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 28 00:18:50.398606 update_engine[1455]: I20260428 00:18:50.398273 1455 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 28 00:18:50.398606 update_engine[1455]: I20260428 00:18:50.398522 1455 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 28 00:18:50.399488 update_engine[1455]: E20260428 00:18:50.399421 1455 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 28 00:18:50.399576 update_engine[1455]: I20260428 00:18:50.399500 1455 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Apr 28 00:18:51.124812 systemd[1]: run-containerd-runc-k8s.io-2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049-runc.zb2VpJ.mount: Deactivated successfully. Apr 28 00:18:54.926011 systemd[1]: Started sshd@22-178.105.21.241:22-50.85.169.122:44998.service - OpenSSH per-connection server daemon (50.85.169.122:44998). Apr 28 00:18:55.058959 sshd[6265]: Accepted publickey for core from 50.85.169.122 port 44998 ssh2: RSA SHA256:0j9rnzg//LrMaH1kTEcAP6LieMSKEVjCW+ZXnbaTdVE Apr 28 00:18:55.061360 sshd[6265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 28 00:18:55.068423 systemd-logind[1454]: New session 22 of user core. Apr 28 00:18:55.077082 systemd[1]: Started session-22.scope - Session 22 of User core. Apr 28 00:18:55.257942 sshd[6265]: pam_unix(sshd:session): session closed for user core Apr 28 00:18:55.264169 systemd[1]: sshd@22-178.105.21.241:22-50.85.169.122:44998.service: Deactivated successfully. Apr 28 00:18:55.266788 systemd[1]: session-22.scope: Deactivated successfully. Apr 28 00:18:55.267964 systemd-logind[1454]: Session 22 logged out. Waiting for processes to exit. Apr 28 00:18:55.269375 systemd-logind[1454]: Removed session 22. Apr 28 00:19:00.402736 update_engine[1455]: I20260428 00:19:00.402282 1455 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 28 00:19:00.403249 update_engine[1455]: I20260428 00:19:00.402755 1455 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 28 00:19:00.403249 update_engine[1455]: I20260428 00:19:00.403090 1455 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 28 00:19:00.404290 update_engine[1455]: E20260428 00:19:00.404218 1455 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 28 00:19:00.404427 update_engine[1455]: I20260428 00:19:00.404309 1455 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 28 00:19:00.404427 update_engine[1455]: I20260428 00:19:00.404320 1455 omaha_request_action.cc:617] Omaha request response: Apr 28 00:19:00.404547 update_engine[1455]: E20260428 00:19:00.404426 1455 omaha_request_action.cc:636] Omaha request network transfer failed. Apr 28 00:19:00.404547 update_engine[1455]: I20260428 00:19:00.404448 1455 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Apr 28 00:19:00.404547 update_engine[1455]: I20260428 00:19:00.404472 1455 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 28 00:19:00.404547 update_engine[1455]: I20260428 00:19:00.404479 1455 update_attempter.cc:306] Processing Done. Apr 28 00:19:00.404547 update_engine[1455]: E20260428 00:19:00.404495 1455 update_attempter.cc:619] Update failed. Apr 28 00:19:00.404547 update_engine[1455]: I20260428 00:19:00.404500 1455 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Apr 28 00:19:00.404547 update_engine[1455]: I20260428 00:19:00.404506 1455 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Apr 28 00:19:00.404547 update_engine[1455]: I20260428 00:19:00.404512 1455 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Apr 28 00:19:00.404850 update_engine[1455]: I20260428 00:19:00.404608 1455 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 28 00:19:00.404850 update_engine[1455]: I20260428 00:19:00.404633 1455 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 28 00:19:00.404850 update_engine[1455]: I20260428 00:19:00.404639 1455 omaha_request_action.cc:272] Request: Apr 28 00:19:00.404850 update_engine[1455]: Apr 28 00:19:00.404850 update_engine[1455]: Apr 28 00:19:00.404850 update_engine[1455]: Apr 28 00:19:00.404850 update_engine[1455]: Apr 28 00:19:00.404850 update_engine[1455]: Apr 28 00:19:00.404850 update_engine[1455]: Apr 28 00:19:00.404850 update_engine[1455]: I20260428 00:19:00.404646 1455 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 28 00:19:00.404850 update_engine[1455]: I20260428 00:19:00.404836 1455 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 28 00:19:00.405177 update_engine[1455]: I20260428 00:19:00.405033 1455 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 28 00:19:00.405583 locksmithd[1488]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Apr 28 00:19:00.406063 update_engine[1455]: E20260428 00:19:00.405928 1455 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 28 00:19:00.406063 update_engine[1455]: I20260428 00:19:00.405992 1455 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 28 00:19:00.406063 update_engine[1455]: I20260428 00:19:00.406000 1455 omaha_request_action.cc:617] Omaha request response: Apr 28 00:19:00.406063 update_engine[1455]: I20260428 00:19:00.406008 1455 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 28 00:19:00.406063 update_engine[1455]: I20260428 00:19:00.406013 1455 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 28 00:19:00.406063 update_engine[1455]: I20260428 00:19:00.406018 1455 update_attempter.cc:306] Processing Done. Apr 28 00:19:00.406063 update_engine[1455]: I20260428 00:19:00.406026 1455 update_attempter.cc:310] Error event sent. Apr 28 00:19:00.406063 update_engine[1455]: I20260428 00:19:00.406037 1455 update_check_scheduler.cc:74] Next update check in 47m4s Apr 28 00:19:00.406558 locksmithd[1488]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Apr 28 00:19:04.579975 systemd[1]: Started sshd@23-178.105.21.241:22-90.226.146.31:37188.service - OpenSSH per-connection server daemon (90.226.146.31:37188). Apr 28 00:19:10.057500 systemd[1]: cri-containerd-e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0.scope: Deactivated successfully. Apr 28 00:19:10.057841 systemd[1]: cri-containerd-e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0.scope: Consumed 14.430s CPU time. Apr 28 00:19:10.081229 containerd[1481]: time="2026-04-28T00:19:10.081147712Z" level=info msg="shim disconnected" id=e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0 namespace=k8s.io Apr 28 00:19:10.081229 containerd[1481]: time="2026-04-28T00:19:10.081214313Z" level=warning msg="cleaning up after shim disconnected" id=e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0 namespace=k8s.io Apr 28 00:19:10.081229 containerd[1481]: time="2026-04-28T00:19:10.081225394Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 28 00:19:10.081843 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0-rootfs.mount: Deactivated successfully. Apr 28 00:19:10.301771 kubelet[2581]: E0428 00:19:10.301715 2581 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:49426->10.0.0.2:2379: read: connection timed out" Apr 28 00:19:10.729269 kubelet[2581]: I0428 00:19:10.729194 2581 scope.go:117] "RemoveContainer" containerID="e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0" Apr 28 00:19:10.736862 containerd[1481]: time="2026-04-28T00:19:10.736802635Z" level=info msg="CreateContainer within sandbox \"b54732ed22708aea8c9222cd6d89520b495adbbd3a7b041ad174a4e96accebb0\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Apr 28 00:19:10.757410 containerd[1481]: time="2026-04-28T00:19:10.757366068Z" level=info msg="CreateContainer within sandbox \"b54732ed22708aea8c9222cd6d89520b495adbbd3a7b041ad174a4e96accebb0\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff\"" Apr 28 00:19:10.758316 containerd[1481]: time="2026-04-28T00:19:10.758289007Z" level=info msg="StartContainer for \"7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff\"" Apr 28 00:19:10.793007 systemd[1]: Started cri-containerd-7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff.scope - libcontainer container 7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff. Apr 28 00:19:10.821663 containerd[1481]: time="2026-04-28T00:19:10.821598580Z" level=info msg="StartContainer for \"7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff\" returns successfully" Apr 28 00:19:11.399293 systemd[1]: cri-containerd-216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5.scope: Deactivated successfully. Apr 28 00:19:11.399929 systemd[1]: cri-containerd-216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5.scope: Consumed 3.897s CPU time, 17.6M memory peak, 0B memory swap peak. Apr 28 00:19:11.424163 containerd[1481]: time="2026-04-28T00:19:11.423536824Z" level=info msg="shim disconnected" id=216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5 namespace=k8s.io Apr 28 00:19:11.424163 containerd[1481]: time="2026-04-28T00:19:11.423612266Z" level=warning msg="cleaning up after shim disconnected" id=216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5 namespace=k8s.io Apr 28 00:19:11.424163 containerd[1481]: time="2026-04-28T00:19:11.423620986Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 28 00:19:11.425815 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5-rootfs.mount: Deactivated successfully. Apr 28 00:19:11.729002 kubelet[2581]: I0428 00:19:11.728525 2581 scope.go:117] "RemoveContainer" containerID="216e39360ab34894191acce4c3a988f08d3f9796ea9870f3d3db72d36a0d2ff5" Apr 28 00:19:11.732948 containerd[1481]: time="2026-04-28T00:19:11.732634373Z" level=info msg="CreateContainer within sandbox \"b758e01b7fed5ba3cfafabab26c0cf47a984f868c9b30445f29b534230d8be15\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Apr 28 00:19:11.746055 containerd[1481]: time="2026-04-28T00:19:11.745985130Z" level=info msg="CreateContainer within sandbox \"b758e01b7fed5ba3cfafabab26c0cf47a984f868c9b30445f29b534230d8be15\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"ff7c23a29bfc68ec1d2e6430a0b20af5c74b816a263569a2ee2e545317fb5643\"" Apr 28 00:19:11.747719 containerd[1481]: time="2026-04-28T00:19:11.746793827Z" level=info msg="StartContainer for \"ff7c23a29bfc68ec1d2e6430a0b20af5c74b816a263569a2ee2e545317fb5643\"" Apr 28 00:19:11.783232 systemd[1]: Started cri-containerd-ff7c23a29bfc68ec1d2e6430a0b20af5c74b816a263569a2ee2e545317fb5643.scope - libcontainer container ff7c23a29bfc68ec1d2e6430a0b20af5c74b816a263569a2ee2e545317fb5643. Apr 28 00:19:11.826120 containerd[1481]: time="2026-04-28T00:19:11.825992194Z" level=info msg="StartContainer for \"ff7c23a29bfc68ec1d2e6430a0b20af5c74b816a263569a2ee2e545317fb5643\" returns successfully" Apr 28 00:19:13.125706 sshd[6340]: Invalid user orangepi from 90.226.146.31 port 37188 Apr 28 00:19:13.819625 sshd[6340]: Connection closed by invalid user orangepi 90.226.146.31 port 37188 [preauth] Apr 28 00:19:13.822917 systemd[1]: sshd@23-178.105.21.241:22-90.226.146.31:37188.service: Deactivated successfully. Apr 28 00:19:15.294718 kubelet[2581]: E0428 00:19:15.286162 2581 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:49256->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4081-3-7-n-51c70c830d.18aa5d42e309b8b5 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4081-3-7-n-51c70c830d,UID:aacf343c9c5f354d3536c75cc0d11c90,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4081-3-7-n-51c70c830d,},FirstTimestamp:2026-04-28 00:19:04.809371829 +0000 UTC m=+199.155466060,LastTimestamp:2026-04-28 00:19:04.809371829 +0000 UTC m=+199.155466060,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-7-n-51c70c830d,}" Apr 28 00:19:16.440267 systemd[1]: cri-containerd-71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000.scope: Deactivated successfully. Apr 28 00:19:16.441170 systemd[1]: cri-containerd-71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000.scope: Consumed 3.349s CPU time, 14.1M memory peak, 0B memory swap peak. Apr 28 00:19:16.467311 containerd[1481]: time="2026-04-28T00:19:16.465553255Z" level=info msg="shim disconnected" id=71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000 namespace=k8s.io Apr 28 00:19:16.467021 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000-rootfs.mount: Deactivated successfully. Apr 28 00:19:16.467832 containerd[1481]: time="2026-04-28T00:19:16.467398771Z" level=warning msg="cleaning up after shim disconnected" id=71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000 namespace=k8s.io Apr 28 00:19:16.467832 containerd[1481]: time="2026-04-28T00:19:16.467423291Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 28 00:19:16.751691 kubelet[2581]: I0428 00:19:16.751330 2581 scope.go:117] "RemoveContainer" containerID="71b8f30e6dc31630821fa035618ea4f03178ec44be3e1466f1896d0646486000" Apr 28 00:19:16.754254 containerd[1481]: time="2026-04-28T00:19:16.754015704Z" level=info msg="CreateContainer within sandbox \"d54450c6df20b4b98a4550cf7e0b028ff4fa4297a3b18073e11264327f670077\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Apr 28 00:19:16.772124 containerd[1481]: time="2026-04-28T00:19:16.772062737Z" level=info msg="CreateContainer within sandbox \"d54450c6df20b4b98a4550cf7e0b028ff4fa4297a3b18073e11264327f670077\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"517b3d73505515aedea911135ace31f0ea1d53398739a2e71acb0f0452caf064\"" Apr 28 00:19:16.775948 containerd[1481]: time="2026-04-28T00:19:16.773310482Z" level=info msg="StartContainer for \"517b3d73505515aedea911135ace31f0ea1d53398739a2e71acb0f0452caf064\"" Apr 28 00:19:16.775556 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2496573303.mount: Deactivated successfully. Apr 28 00:19:16.810004 systemd[1]: Started cri-containerd-517b3d73505515aedea911135ace31f0ea1d53398739a2e71acb0f0452caf064.scope - libcontainer container 517b3d73505515aedea911135ace31f0ea1d53398739a2e71acb0f0452caf064. Apr 28 00:19:16.843652 containerd[1481]: time="2026-04-28T00:19:16.843603778Z" level=info msg="StartContainer for \"517b3d73505515aedea911135ace31f0ea1d53398739a2e71acb0f0452caf064\" returns successfully" Apr 28 00:19:20.302511 kubelet[2581]: E0428 00:19:20.302316 2581 controller.go:195] "Failed to update lease" err="Put \"https://178.105.21.241:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-7-n-51c70c830d?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 28 00:19:21.120440 systemd[1]: run-containerd-runc-k8s.io-2b4c51cd9b0a438841e2ca48378f0223fef6f33be28f516be8e4421de3138049-runc.Y8kmOw.mount: Deactivated successfully. Apr 28 00:19:21.174202 kubelet[2581]: I0428 00:19:21.174140 2581 status_manager.go:895] "Failed to get status for pod" podUID="d21fdc26-3471-477d-a2b1-baa19a57994e" pod="tigera-operator/tigera-operator-8458958b4d-psz5j" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:49352->10.0.0.2:2379: read: connection timed out" Apr 28 00:19:22.108874 systemd[1]: cri-containerd-7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff.scope: Deactivated successfully. Apr 28 00:19:22.147821 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff-rootfs.mount: Deactivated successfully. Apr 28 00:19:22.151780 containerd[1481]: time="2026-04-28T00:19:22.151710966Z" level=info msg="shim disconnected" id=7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff namespace=k8s.io Apr 28 00:19:22.151780 containerd[1481]: time="2026-04-28T00:19:22.151769447Z" level=warning msg="cleaning up after shim disconnected" id=7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff namespace=k8s.io Apr 28 00:19:22.151780 containerd[1481]: time="2026-04-28T00:19:22.151781767Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 28 00:19:22.772758 kubelet[2581]: I0428 00:19:22.772720 2581 scope.go:117] "RemoveContainer" containerID="e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0" Apr 28 00:19:22.773240 kubelet[2581]: I0428 00:19:22.773010 2581 scope.go:117] "RemoveContainer" containerID="7fbae89fee596df91e004c1d0b10d2eec852b57f2609d35fd237ae4a4aecb5ff" Apr 28 00:19:22.773240 kubelet[2581]: E0428 00:19:22.773134 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tigera-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=tigera-operator pod=tigera-operator-8458958b4d-psz5j_tigera-operator(d21fdc26-3471-477d-a2b1-baa19a57994e)\"" pod="tigera-operator/tigera-operator-8458958b4d-psz5j" podUID="d21fdc26-3471-477d-a2b1-baa19a57994e" Apr 28 00:19:22.775003 containerd[1481]: time="2026-04-28T00:19:22.774646172Z" level=info msg="RemoveContainer for \"e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0\"" Apr 28 00:19:22.779286 containerd[1481]: time="2026-04-28T00:19:22.779222816Z" level=info msg="RemoveContainer for \"e154f99f01fd16eca154b27f77920f7e701e0f138033ab200ce0b8f276efedd0\" returns successfully"