Mar 7 01:48:11.171208 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Mar 7 01:48:11.171228 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Fri Mar 6 22:59:59 -00 2026 Mar 7 01:48:11.171236 kernel: KASLR enabled Mar 7 01:48:11.171242 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Mar 7 01:48:11.171249 kernel: printk: bootconsole [pl11] enabled Mar 7 01:48:11.171255 kernel: efi: EFI v2.7 by EDK II Mar 7 01:48:11.171262 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3f215018 RNG=0x3fd5f998 MEMRESERVE=0x3e44ee18 Mar 7 01:48:11.171268 kernel: random: crng init done Mar 7 01:48:11.171274 kernel: ACPI: Early table checksum verification disabled Mar 7 01:48:11.171280 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) Mar 7 01:48:11.171286 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171292 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171300 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) Mar 7 01:48:11.171306 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171313 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171320 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171326 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171334 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171340 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171347 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Mar 7 01:48:11.171353 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171360 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Mar 7 01:48:11.171366 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Mar 7 01:48:11.171372 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] Mar 7 01:48:11.171379 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] Mar 7 01:48:11.171385 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] Mar 7 01:48:11.171392 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] Mar 7 01:48:11.171398 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] Mar 7 01:48:11.171406 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] Mar 7 01:48:11.171413 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] Mar 7 01:48:11.171419 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] Mar 7 01:48:11.171425 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] Mar 7 01:48:11.171432 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] Mar 7 01:48:11.171438 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] Mar 7 01:48:11.171444 kernel: NUMA: NODE_DATA [mem 0x1bf7ef800-0x1bf7f4fff] Mar 7 01:48:11.171451 kernel: Zone ranges: Mar 7 01:48:11.171457 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Mar 7 01:48:11.171463 kernel: DMA32 empty Mar 7 01:48:11.171470 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Mar 7 01:48:11.171476 kernel: Movable zone start for each node Mar 7 01:48:11.171487 kernel: Early memory node ranges Mar 7 01:48:11.171494 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Mar 7 01:48:11.171501 kernel: node 0: [mem 0x0000000000824000-0x000000003e54ffff] Mar 7 01:48:11.171507 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] Mar 7 01:48:11.171514 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] Mar 7 01:48:11.171522 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] Mar 7 01:48:11.171529 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] Mar 7 01:48:11.171536 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Mar 7 01:48:11.171543 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Mar 7 01:48:11.171550 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Mar 7 01:48:11.171556 kernel: psci: probing for conduit method from ACPI. Mar 7 01:48:11.171563 kernel: psci: PSCIv1.1 detected in firmware. Mar 7 01:48:11.171570 kernel: psci: Using standard PSCI v0.2 function IDs Mar 7 01:48:11.171577 kernel: psci: MIGRATE_INFO_TYPE not supported. Mar 7 01:48:11.171584 kernel: psci: SMC Calling Convention v1.4 Mar 7 01:48:11.171591 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Mar 7 01:48:11.171598 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Mar 7 01:48:11.171607 kernel: percpu: Embedded 30 pages/cpu s85736 r8192 d28952 u122880 Mar 7 01:48:11.171614 kernel: pcpu-alloc: s85736 r8192 d28952 u122880 alloc=30*4096 Mar 7 01:48:11.171620 kernel: pcpu-alloc: [0] 0 [0] 1 Mar 7 01:48:11.171627 kernel: Detected PIPT I-cache on CPU0 Mar 7 01:48:11.171634 kernel: CPU features: detected: GIC system register CPU interface Mar 7 01:48:11.171641 kernel: CPU features: detected: Hardware dirty bit management Mar 7 01:48:11.171648 kernel: CPU features: detected: Spectre-BHB Mar 7 01:48:11.171655 kernel: CPU features: kernel page table isolation forced ON by KASLR Mar 7 01:48:11.171662 kernel: CPU features: detected: Kernel page table isolation (KPTI) Mar 7 01:48:11.171668 kernel: CPU features: detected: ARM erratum 1418040 Mar 7 01:48:11.171675 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Mar 7 01:48:11.171683 kernel: CPU features: detected: SSBS not fully self-synchronizing Mar 7 01:48:11.171690 kernel: alternatives: applying boot alternatives Mar 7 01:48:11.171698 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=9d22c40559a0d209dc0fcc2dfdd5ddf9671e6da0cc59463f610ba522f01325a6 Mar 7 01:48:11.171705 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 7 01:48:11.171712 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 7 01:48:11.171719 kernel: Fallback order for Node 0: 0 Mar 7 01:48:11.171726 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Mar 7 01:48:11.171732 kernel: Policy zone: Normal Mar 7 01:48:11.171739 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 7 01:48:11.171746 kernel: software IO TLB: area num 2. Mar 7 01:48:11.171753 kernel: software IO TLB: mapped [mem 0x000000003a44e000-0x000000003e44e000] (64MB) Mar 7 01:48:11.171761 kernel: Memory: 3982636K/4194160K available (10304K kernel code, 2180K rwdata, 8116K rodata, 39424K init, 897K bss, 211524K reserved, 0K cma-reserved) Mar 7 01:48:11.171768 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 7 01:48:11.171775 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 7 01:48:11.171782 kernel: rcu: RCU event tracing is enabled. Mar 7 01:48:11.171790 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 7 01:48:11.171796 kernel: Trampoline variant of Tasks RCU enabled. Mar 7 01:48:11.171804 kernel: Tracing variant of Tasks RCU enabled. Mar 7 01:48:11.171811 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 7 01:48:11.171818 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 7 01:48:11.171825 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 7 01:48:11.171832 kernel: GICv3: 960 SPIs implemented Mar 7 01:48:11.171841 kernel: GICv3: 0 Extended SPIs implemented Mar 7 01:48:11.171848 kernel: Root IRQ handler: gic_handle_irq Mar 7 01:48:11.171854 kernel: GICv3: GICv3 features: 16 PPIs, RSS Mar 7 01:48:11.171861 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Mar 7 01:48:11.171868 kernel: ITS: No ITS available, not enabling LPIs Mar 7 01:48:11.171875 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 7 01:48:11.171882 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 7 01:48:11.171889 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Mar 7 01:48:11.171896 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Mar 7 01:48:11.171903 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Mar 7 01:48:11.171910 kernel: Console: colour dummy device 80x25 Mar 7 01:48:11.171918 kernel: printk: console [tty1] enabled Mar 7 01:48:11.171925 kernel: ACPI: Core revision 20230628 Mar 7 01:48:11.171933 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Mar 7 01:48:11.171940 kernel: pid_max: default: 32768 minimum: 301 Mar 7 01:48:11.171947 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 7 01:48:11.171954 kernel: landlock: Up and running. Mar 7 01:48:11.171961 kernel: SELinux: Initializing. Mar 7 01:48:11.171968 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.171975 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.171984 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 01:48:11.171991 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 01:48:11.171998 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0x100000e, misc 0x31e1 Mar 7 01:48:11.172005 kernel: Hyper-V: Host Build 10.0.26100.1480-1-0 Mar 7 01:48:11.172012 kernel: Hyper-V: enabling crash_kexec_post_notifiers Mar 7 01:48:11.172019 kernel: rcu: Hierarchical SRCU implementation. Mar 7 01:48:11.172026 kernel: rcu: Max phase no-delay instances is 400. Mar 7 01:48:11.172034 kernel: Remapping and enabling EFI services. Mar 7 01:48:11.172046 kernel: smp: Bringing up secondary CPUs ... Mar 7 01:48:11.172054 kernel: Detected PIPT I-cache on CPU1 Mar 7 01:48:11.172061 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Mar 7 01:48:11.172075 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 7 01:48:11.172085 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Mar 7 01:48:11.172092 kernel: smp: Brought up 1 node, 2 CPUs Mar 7 01:48:11.172100 kernel: SMP: Total of 2 processors activated. Mar 7 01:48:11.172108 kernel: CPU features: detected: 32-bit EL0 Support Mar 7 01:48:11.172115 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Mar 7 01:48:11.172124 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Mar 7 01:48:11.172131 kernel: CPU features: detected: CRC32 instructions Mar 7 01:48:11.172139 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Mar 7 01:48:11.172146 kernel: CPU features: detected: LSE atomic instructions Mar 7 01:48:11.172154 kernel: CPU features: detected: Privileged Access Never Mar 7 01:48:11.172161 kernel: CPU: All CPU(s) started at EL1 Mar 7 01:48:11.172168 kernel: alternatives: applying system-wide alternatives Mar 7 01:48:11.172175 kernel: devtmpfs: initialized Mar 7 01:48:11.172183 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 7 01:48:11.172192 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 7 01:48:11.172199 kernel: pinctrl core: initialized pinctrl subsystem Mar 7 01:48:11.172206 kernel: SMBIOS 3.1.0 present. Mar 7 01:48:11.172214 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 Mar 7 01:48:11.172221 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 7 01:48:11.172229 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 7 01:48:11.172236 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 7 01:48:11.172244 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 7 01:48:11.172251 kernel: audit: initializing netlink subsys (disabled) Mar 7 01:48:11.172260 kernel: audit: type=2000 audit(0.047:1): state=initialized audit_enabled=0 res=1 Mar 7 01:48:11.172268 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 7 01:48:11.172275 kernel: cpuidle: using governor menu Mar 7 01:48:11.172283 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 7 01:48:11.172291 kernel: ASID allocator initialised with 32768 entries Mar 7 01:48:11.172299 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 7 01:48:11.172306 kernel: Serial: AMBA PL011 UART driver Mar 7 01:48:11.172314 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Mar 7 01:48:11.172321 kernel: Modules: 0 pages in range for non-PLT usage Mar 7 01:48:11.172330 kernel: Modules: 509008 pages in range for PLT usage Mar 7 01:48:11.172338 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 7 01:48:11.172346 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Mar 7 01:48:11.172353 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Mar 7 01:48:11.172361 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Mar 7 01:48:11.172368 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 7 01:48:11.172375 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Mar 7 01:48:11.172383 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Mar 7 01:48:11.172390 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Mar 7 01:48:11.172399 kernel: ACPI: Added _OSI(Module Device) Mar 7 01:48:11.172406 kernel: ACPI: Added _OSI(Processor Device) Mar 7 01:48:11.172413 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 7 01:48:11.172420 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 7 01:48:11.172428 kernel: ACPI: Interpreter enabled Mar 7 01:48:11.172435 kernel: ACPI: Using GIC for interrupt routing Mar 7 01:48:11.172443 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Mar 7 01:48:11.172450 kernel: printk: console [ttyAMA0] enabled Mar 7 01:48:11.172457 kernel: printk: bootconsole [pl11] disabled Mar 7 01:48:11.172466 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Mar 7 01:48:11.172474 kernel: iommu: Default domain type: Translated Mar 7 01:48:11.172481 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 7 01:48:11.172488 kernel: efivars: Registered efivars operations Mar 7 01:48:11.172495 kernel: vgaarb: loaded Mar 7 01:48:11.172503 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 7 01:48:11.172510 kernel: VFS: Disk quotas dquot_6.6.0 Mar 7 01:48:11.172517 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 7 01:48:11.172525 kernel: pnp: PnP ACPI init Mar 7 01:48:11.172533 kernel: pnp: PnP ACPI: found 0 devices Mar 7 01:48:11.172540 kernel: NET: Registered PF_INET protocol family Mar 7 01:48:11.172548 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 7 01:48:11.172555 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 7 01:48:11.172563 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 7 01:48:11.172570 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 7 01:48:11.172577 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 7 01:48:11.172585 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 7 01:48:11.172592 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.172601 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.172608 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 7 01:48:11.172616 kernel: PCI: CLS 0 bytes, default 64 Mar 7 01:48:11.172623 kernel: kvm [1]: HYP mode not available Mar 7 01:48:11.172630 kernel: Initialise system trusted keyrings Mar 7 01:48:11.172638 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 7 01:48:11.172645 kernel: Key type asymmetric registered Mar 7 01:48:11.172652 kernel: Asymmetric key parser 'x509' registered Mar 7 01:48:11.172659 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 7 01:48:11.172668 kernel: io scheduler mq-deadline registered Mar 7 01:48:11.172676 kernel: io scheduler kyber registered Mar 7 01:48:11.172683 kernel: io scheduler bfq registered Mar 7 01:48:11.172690 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 7 01:48:11.172697 kernel: thunder_xcv, ver 1.0 Mar 7 01:48:11.172705 kernel: thunder_bgx, ver 1.0 Mar 7 01:48:11.172712 kernel: nicpf, ver 1.0 Mar 7 01:48:11.172719 kernel: nicvf, ver 1.0 Mar 7 01:48:11.172837 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 7 01:48:11.172916 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-03-07T01:48:10 UTC (1772848090) Mar 7 01:48:11.172926 kernel: efifb: probing for efifb Mar 7 01:48:11.172934 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Mar 7 01:48:11.172941 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Mar 7 01:48:11.172949 kernel: efifb: scrolling: redraw Mar 7 01:48:11.172956 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 7 01:48:11.172963 kernel: Console: switching to colour frame buffer device 128x48 Mar 7 01:48:11.172971 kernel: fb0: EFI VGA frame buffer device Mar 7 01:48:11.172980 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Mar 7 01:48:11.172989 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 7 01:48:11.172996 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 6 counters available Mar 7 01:48:11.173004 kernel: watchdog: Delayed init of the lockup detector failed: -19 Mar 7 01:48:11.173011 kernel: watchdog: Hard watchdog permanently disabled Mar 7 01:48:11.173018 kernel: NET: Registered PF_INET6 protocol family Mar 7 01:48:11.173025 kernel: Segment Routing with IPv6 Mar 7 01:48:11.173033 kernel: In-situ OAM (IOAM) with IPv6 Mar 7 01:48:11.173040 kernel: NET: Registered PF_PACKET protocol family Mar 7 01:48:11.173049 kernel: Key type dns_resolver registered Mar 7 01:48:11.173056 kernel: registered taskstats version 1 Mar 7 01:48:11.173068 kernel: Loading compiled-in X.509 certificates Mar 7 01:48:11.173077 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: e62b4e4ebcb406beff1271ecc7444548c4ab67e9' Mar 7 01:48:11.173084 kernel: Key type .fscrypt registered Mar 7 01:48:11.173091 kernel: Key type fscrypt-provisioning registered Mar 7 01:48:11.173098 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 7 01:48:11.173106 kernel: ima: Allocated hash algorithm: sha1 Mar 7 01:48:11.173113 kernel: ima: No architecture policies found Mar 7 01:48:11.173122 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 7 01:48:11.173130 kernel: clk: Disabling unused clocks Mar 7 01:48:11.173137 kernel: Freeing unused kernel memory: 39424K Mar 7 01:48:11.173144 kernel: Run /init as init process Mar 7 01:48:11.173151 kernel: with arguments: Mar 7 01:48:11.173158 kernel: /init Mar 7 01:48:11.173165 kernel: with environment: Mar 7 01:48:11.173173 kernel: HOME=/ Mar 7 01:48:11.173180 kernel: TERM=linux Mar 7 01:48:11.173189 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 7 01:48:11.173200 systemd[1]: Detected virtualization microsoft. Mar 7 01:48:11.173208 systemd[1]: Detected architecture arm64. Mar 7 01:48:11.173216 systemd[1]: Running in initrd. Mar 7 01:48:11.173224 systemd[1]: No hostname configured, using default hostname. Mar 7 01:48:11.173231 systemd[1]: Hostname set to . Mar 7 01:48:11.173239 systemd[1]: Initializing machine ID from random generator. Mar 7 01:48:11.173247 systemd[1]: Queued start job for default target initrd.target. Mar 7 01:48:11.173257 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:48:11.173265 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:48:11.173274 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 7 01:48:11.173282 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 7 01:48:11.173290 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 7 01:48:11.173298 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 7 01:48:11.173307 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 7 01:48:11.173317 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 7 01:48:11.173325 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:48:11.173333 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:48:11.173343 systemd[1]: Reached target paths.target - Path Units. Mar 7 01:48:11.173351 systemd[1]: Reached target slices.target - Slice Units. Mar 7 01:48:11.173359 systemd[1]: Reached target swap.target - Swaps. Mar 7 01:48:11.173367 systemd[1]: Reached target timers.target - Timer Units. Mar 7 01:48:11.173375 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 01:48:11.173385 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 01:48:11.173393 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 7 01:48:11.173401 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 7 01:48:11.173409 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:48:11.173417 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 7 01:48:11.173424 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:48:11.173432 systemd[1]: Reached target sockets.target - Socket Units. Mar 7 01:48:11.173440 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 7 01:48:11.173450 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 7 01:48:11.173457 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 7 01:48:11.173465 systemd[1]: Starting systemd-fsck-usr.service... Mar 7 01:48:11.173473 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 7 01:48:11.173496 systemd-journald[218]: Collecting audit messages is disabled. Mar 7 01:48:11.173516 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 7 01:48:11.173524 systemd-journald[218]: Journal started Mar 7 01:48:11.173542 systemd-journald[218]: Runtime Journal (/run/log/journal/9be137b8f6ae4d63975e41ca0ff2e618) is 8.0M, max 78.5M, 70.5M free. Mar 7 01:48:11.184997 systemd-modules-load[219]: Inserted module 'overlay' Mar 7 01:48:11.189739 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:11.204468 systemd[1]: Started systemd-journald.service - Journal Service. Mar 7 01:48:11.205434 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 7 01:48:11.227157 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 7 01:48:11.227186 kernel: Bridge firewalling registered Mar 7 01:48:11.218048 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:48:11.226140 systemd-modules-load[219]: Inserted module 'br_netfilter' Mar 7 01:48:11.232084 systemd[1]: Finished systemd-fsck-usr.service. Mar 7 01:48:11.239099 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 7 01:48:11.247278 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:11.269323 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:11.281669 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 01:48:11.297288 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 7 01:48:11.313243 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 7 01:48:11.320860 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:11.333585 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:48:11.350161 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 7 01:48:11.356119 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:48:11.375267 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 7 01:48:11.381237 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 7 01:48:11.403881 dracut-cmdline[253]: dracut-dracut-053 Mar 7 01:48:11.405207 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 7 01:48:11.425391 dracut-cmdline[253]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=9d22c40559a0d209dc0fcc2dfdd5ddf9671e6da0cc59463f610ba522f01325a6 Mar 7 01:48:11.417609 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:48:11.430952 systemd-resolved[254]: Positive Trust Anchors: Mar 7 01:48:11.430962 systemd-resolved[254]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 7 01:48:11.430994 systemd-resolved[254]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 7 01:48:11.433865 systemd-resolved[254]: Defaulting to hostname 'linux'. Mar 7 01:48:11.461475 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 7 01:48:11.469530 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:48:11.574089 kernel: SCSI subsystem initialized Mar 7 01:48:11.583079 kernel: Loading iSCSI transport class v2.0-870. Mar 7 01:48:11.591094 kernel: iscsi: registered transport (tcp) Mar 7 01:48:11.606888 kernel: iscsi: registered transport (qla4xxx) Mar 7 01:48:11.606929 kernel: QLogic iSCSI HBA Driver Mar 7 01:48:11.645115 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 7 01:48:11.655362 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 7 01:48:11.683865 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 7 01:48:11.683923 kernel: device-mapper: uevent: version 1.0.3 Mar 7 01:48:11.689123 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 7 01:48:11.738093 kernel: raid6: neonx8 gen() 15785 MB/s Mar 7 01:48:11.755080 kernel: raid6: neonx4 gen() 15694 MB/s Mar 7 01:48:11.774072 kernel: raid6: neonx2 gen() 13277 MB/s Mar 7 01:48:11.796089 kernel: raid6: neonx1 gen() 10538 MB/s Mar 7 01:48:11.811076 kernel: raid6: int64x8 gen() 6580 MB/s Mar 7 01:48:11.830072 kernel: raid6: int64x4 gen() 7371 MB/s Mar 7 01:48:11.850074 kernel: raid6: int64x2 gen() 6146 MB/s Mar 7 01:48:11.872674 kernel: raid6: int64x1 gen() 5071 MB/s Mar 7 01:48:11.872695 kernel: raid6: using algorithm neonx8 gen() 15785 MB/s Mar 7 01:48:11.894440 kernel: raid6: .... xor() 12041 MB/s, rmw enabled Mar 7 01:48:11.894460 kernel: raid6: using neon recovery algorithm Mar 7 01:48:11.905191 kernel: xor: measuring software checksum speed Mar 7 01:48:11.905221 kernel: 8regs : 19797 MB/sec Mar 7 01:48:11.907912 kernel: 32regs : 19688 MB/sec Mar 7 01:48:11.910621 kernel: arm64_neon : 27105 MB/sec Mar 7 01:48:11.913795 kernel: xor: using function: arm64_neon (27105 MB/sec) Mar 7 01:48:11.964091 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 7 01:48:11.974452 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 7 01:48:11.989216 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:48:12.009302 systemd-udevd[439]: Using default interface naming scheme 'v255'. Mar 7 01:48:12.013731 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:48:12.030212 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 7 01:48:12.044365 dracut-pre-trigger[452]: rd.md=0: removing MD RAID activation Mar 7 01:48:12.071625 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 01:48:12.083515 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 7 01:48:12.120835 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:48:12.136221 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 7 01:48:12.157159 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 7 01:48:12.172348 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 01:48:12.180673 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:48:12.196195 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 7 01:48:12.219647 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 7 01:48:12.237303 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:48:12.247317 kernel: hv_vmbus: Vmbus version:5.3 Mar 7 01:48:12.237413 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:12.275792 kernel: hv_vmbus: registering driver hyperv_keyboard Mar 7 01:48:12.275814 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Mar 7 01:48:12.275492 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:12.291114 kernel: pps_core: LinuxPPS API ver. 1 registered Mar 7 01:48:12.282405 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:48:12.310418 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Mar 7 01:48:12.310441 kernel: PTP clock support registered Mar 7 01:48:12.282564 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.303334 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.332382 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.360718 kernel: hv_vmbus: registering driver hid_hyperv Mar 7 01:48:12.360743 kernel: hv_utils: Registering HyperV Utility Driver Mar 7 01:48:12.360753 kernel: hv_vmbus: registering driver hv_utils Mar 7 01:48:12.360763 kernel: hv_vmbus: registering driver hv_storvsc Mar 7 01:48:12.360773 kernel: hv_utils: Heartbeat IC version 3.0 Mar 7 01:48:12.360782 kernel: hv_utils: Shutdown IC version 3.2 Mar 7 01:48:12.340458 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 7 01:48:11.970378 kernel: hv_utils: TimeSync IC version 4.0 Mar 7 01:48:11.981887 kernel: hv_vmbus: registering driver hv_netvsc Mar 7 01:48:11.981904 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Mar 7 01:48:11.981913 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Mar 7 01:48:11.982028 systemd-journald[218]: Time jumped backwards, rotating. Mar 7 01:48:11.958860 systemd-resolved[254]: Clock change detected. Flushing caches. Mar 7 01:48:12.003302 kernel: scsi host0: storvsc_host_t Mar 7 01:48:12.009316 kernel: scsi host1: storvsc_host_t Mar 7 01:48:12.009507 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Mar 7 01:48:12.009543 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 5 Mar 7 01:48:11.960883 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.003912 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:12.027171 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:48:12.027305 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:12.039856 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:48:12.072201 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Mar 7 01:48:12.072388 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: VF slot 1 added Mar 7 01:48:12.072502 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 7 01:48:12.039920 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.047003 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.099649 kernel: hv_vmbus: registering driver hv_pci Mar 7 01:48:12.099671 kernel: hv_pci d7ed302e-4556-46d5-ba91-8abdbb0cefa4: PCI VMBus probing: Using version 0x10004 Mar 7 01:48:12.079703 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.110110 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Mar 7 01:48:12.110300 kernel: hv_pci d7ed302e-4556-46d5-ba91-8abdbb0cefa4: PCI host bridge to bus 4556:00 Mar 7 01:48:12.118410 kernel: pci_bus 4556:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Mar 7 01:48:12.129181 kernel: pci_bus 4556:00: No busn resource found for root bus, will use [bus 00-ff] Mar 7 01:48:12.129359 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Mar 7 01:48:12.129487 kernel: pci 4556:00:02.0: [15b3:1018] type 00 class 0x020000 Mar 7 01:48:12.137150 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Mar 7 01:48:12.137660 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.163031 kernel: sd 0:0:0:0: [sda] Write Protect is off Mar 7 01:48:12.163199 kernel: pci 4556:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Mar 7 01:48:12.163220 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Mar 7 01:48:12.163309 kernel: pci 4556:00:02.0: enabling Extended Tags Mar 7 01:48:12.163323 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Mar 7 01:48:12.165669 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:12.213143 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 01:48:12.213166 kernel: pci 4556:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 4556:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Mar 7 01:48:12.213346 kernel: pci_bus 4556:00: busn_res: [bus 00-ff] end is updated to 00 Mar 7 01:48:12.213466 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#218 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 7 01:48:12.213574 kernel: pci 4556:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Mar 7 01:48:12.213666 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Mar 7 01:48:12.224324 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:12.258894 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#255 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 7 01:48:12.268353 kernel: mlx5_core 4556:00:02.0: enabling device (0000 -> 0002) Mar 7 01:48:12.274513 kernel: mlx5_core 4556:00:02.0: firmware version: 16.30.5026 Mar 7 01:48:12.475243 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: VF registering: eth1 Mar 7 01:48:12.475431 kernel: mlx5_core 4556:00:02.0 eth1: joined to eth0 Mar 7 01:48:12.480505 kernel: mlx5_core 4556:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Mar 7 01:48:12.490466 kernel: mlx5_core 4556:00:02.0 enP17750s1: renamed from eth1 Mar 7 01:48:12.901461 kernel: BTRFS: device fsid 237c8587-8110-47ef-99f9-37e4ed4d3b31 devid 1 transid 36 /dev/sda3 scanned by (udev-worker) (488) Mar 7 01:48:12.914815 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Mar 7 01:48:12.920072 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Mar 7 01:48:12.952618 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 7 01:48:13.969912 disk-uuid[597]: Warning: The kernel is still using the old partition table. Mar 7 01:48:13.969912 disk-uuid[597]: The new table will be used at the next reboot or after you Mar 7 01:48:13.969912 disk-uuid[597]: run partprobe(8) or kpartx(8) Mar 7 01:48:13.969912 disk-uuid[597]: The operation has completed successfully. Mar 7 01:48:14.848642 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by (udev-worker) (512) Mar 7 01:48:14.863055 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Mar 7 01:48:15.039137 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Mar 7 01:48:15.358722 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Mar 7 01:48:15.828505 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 7 01:48:15.828614 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 7 01:48:15.846574 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 7 01:48:15.856478 sh[699]: Success Mar 7 01:48:15.887510 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Mar 7 01:48:16.178336 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 7 01:48:16.183016 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 7 01:48:16.195572 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 7 01:48:16.230009 kernel: BTRFS info (device dm-0): first mount of filesystem 237c8587-8110-47ef-99f9-37e4ed4d3b31 Mar 7 01:48:16.230061 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:16.235380 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 7 01:48:16.239375 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 7 01:48:16.242657 kernel: BTRFS info (device dm-0): using free space tree Mar 7 01:48:16.582941 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 7 01:48:16.587212 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 7 01:48:16.608748 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 7 01:48:16.614589 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 7 01:48:16.648644 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:16.648698 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:16.652118 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:48:16.693472 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:48:16.702291 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 7 01:48:16.711938 kernel: BTRFS info (device sda6): last unmount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:16.719250 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 7 01:48:16.732684 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 7 01:48:16.754327 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 01:48:16.769558 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 7 01:48:16.793733 systemd-networkd[883]: lo: Link UP Mar 7 01:48:16.793742 systemd-networkd[883]: lo: Gained carrier Mar 7 01:48:16.795386 systemd-networkd[883]: Enumeration completed Mar 7 01:48:16.795923 systemd-networkd[883]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:48:16.795926 systemd-networkd[883]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:48:16.800092 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 7 01:48:16.805381 systemd[1]: Reached target network.target - Network. Mar 7 01:48:16.876540 kernel: mlx5_core 4556:00:02.0 enP17750s1: Link up Mar 7 01:48:16.916719 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: Data path switched to VF: enP17750s1 Mar 7 01:48:16.917240 systemd-networkd[883]: enP17750s1: Link UP Mar 7 01:48:16.917340 systemd-networkd[883]: eth0: Link UP Mar 7 01:48:16.917503 systemd-networkd[883]: eth0: Gained carrier Mar 7 01:48:16.917512 systemd-networkd[883]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:48:16.926615 systemd-networkd[883]: enP17750s1: Gained carrier Mar 7 01:48:16.945488 systemd-networkd[883]: eth0: DHCPv4 address 10.200.20.15/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 7 01:48:17.565917 ignition[863]: Ignition 2.19.0 Mar 7 01:48:17.565928 ignition[863]: Stage: fetch-offline Mar 7 01:48:17.570058 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 01:48:17.565964 ignition[863]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.565972 ignition[863]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.566062 ignition[863]: parsed url from cmdline: "" Mar 7 01:48:17.566064 ignition[863]: no config URL provided Mar 7 01:48:17.566069 ignition[863]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 01:48:17.591708 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 7 01:48:17.566076 ignition[863]: no config at "/usr/lib/ignition/user.ign" Mar 7 01:48:17.566080 ignition[863]: failed to fetch config: resource requires networking Mar 7 01:48:17.566410 ignition[863]: Ignition finished successfully Mar 7 01:48:17.610690 ignition[893]: Ignition 2.19.0 Mar 7 01:48:17.610697 ignition[893]: Stage: fetch Mar 7 01:48:17.610889 ignition[893]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.610898 ignition[893]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.610999 ignition[893]: parsed url from cmdline: "" Mar 7 01:48:17.611002 ignition[893]: no config URL provided Mar 7 01:48:17.611006 ignition[893]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 01:48:17.611014 ignition[893]: no config at "/usr/lib/ignition/user.ign" Mar 7 01:48:17.611036 ignition[893]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Mar 7 01:48:17.697435 ignition[893]: GET result: OK Mar 7 01:48:17.697532 ignition[893]: config has been read from IMDS userdata Mar 7 01:48:17.697579 ignition[893]: parsing config with SHA512: 91949a450c5096514d3c2c8c0d84e48291905018ba2b69f9b8c000712f659a4ad5d8dfc05d829a160b79aff16badf2df05c541342e2d09b31f0c75a1fc715f3c Mar 7 01:48:17.701744 unknown[893]: fetched base config from "system" Mar 7 01:48:17.702160 ignition[893]: fetch: fetch complete Mar 7 01:48:17.701751 unknown[893]: fetched base config from "system" Mar 7 01:48:17.702165 ignition[893]: fetch: fetch passed Mar 7 01:48:17.701756 unknown[893]: fetched user config from "azure" Mar 7 01:48:17.702204 ignition[893]: Ignition finished successfully Mar 7 01:48:17.705492 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 7 01:48:17.726629 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 7 01:48:17.744131 ignition[899]: Ignition 2.19.0 Mar 7 01:48:17.744141 ignition[899]: Stage: kargs Mar 7 01:48:17.744316 ignition[899]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.749608 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 7 01:48:17.744328 ignition[899]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.745409 ignition[899]: kargs: kargs passed Mar 7 01:48:17.745468 ignition[899]: Ignition finished successfully Mar 7 01:48:17.773000 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 7 01:48:17.788280 ignition[906]: Ignition 2.19.0 Mar 7 01:48:17.788289 ignition[906]: Stage: disks Mar 7 01:48:17.792096 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 7 01:48:17.788463 ignition[906]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.798663 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 7 01:48:17.788474 ignition[906]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.806764 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 7 01:48:17.789375 ignition[906]: disks: disks passed Mar 7 01:48:17.815170 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 7 01:48:17.789422 ignition[906]: Ignition finished successfully Mar 7 01:48:17.823596 systemd[1]: Reached target sysinit.target - System Initialization. Mar 7 01:48:17.832140 systemd[1]: Reached target basic.target - Basic System. Mar 7 01:48:17.850728 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 7 01:48:17.927014 systemd-fsck[914]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Mar 7 01:48:17.934709 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 7 01:48:17.948625 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 7 01:48:18.003455 kernel: EXT4-fs (sda9): mounted filesystem 596a8ea8-9d3d-4d06-a56e-9d3ebd3cb76d r/w with ordered data mode. Quota mode: none. Mar 7 01:48:18.004470 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 7 01:48:18.008032 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 7 01:48:18.060513 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 01:48:18.078459 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (925) Mar 7 01:48:18.089303 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:18.089357 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:18.092603 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:48:18.099455 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:48:18.100564 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 7 01:48:18.106615 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 7 01:48:18.113697 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 7 01:48:18.113731 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 01:48:18.124861 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 01:48:18.138269 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 7 01:48:18.154773 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 7 01:48:18.675477 coreos-metadata[942]: Mar 07 01:48:18.675 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 7 01:48:18.682708 coreos-metadata[942]: Mar 07 01:48:18.682 INFO Fetch successful Mar 7 01:48:18.686800 coreos-metadata[942]: Mar 07 01:48:18.686 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Mar 7 01:48:18.695240 systemd-networkd[883]: eth0: Gained IPv6LL Mar 7 01:48:18.705164 coreos-metadata[942]: Mar 07 01:48:18.705 INFO Fetch successful Mar 7 01:48:18.721984 coreos-metadata[942]: Mar 07 01:48:18.720 INFO wrote hostname ci-4081.3.6-n-e8b1fc37af to /sysroot/etc/hostname Mar 7 01:48:18.729407 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 01:48:18.962338 initrd-setup-root[955]: cut: /sysroot/etc/passwd: No such file or directory Mar 7 01:48:19.017719 initrd-setup-root[962]: cut: /sysroot/etc/group: No such file or directory Mar 7 01:48:19.039450 initrd-setup-root[969]: cut: /sysroot/etc/shadow: No such file or directory Mar 7 01:48:19.044813 initrd-setup-root[976]: cut: /sysroot/etc/gshadow: No such file or directory Mar 7 01:48:20.269426 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 7 01:48:20.281623 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 7 01:48:20.288583 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 7 01:48:20.305802 kernel: BTRFS info (device sda6): last unmount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:20.306332 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 7 01:48:20.327671 ignition[1043]: INFO : Ignition 2.19.0 Mar 7 01:48:20.331005 ignition[1043]: INFO : Stage: mount Mar 7 01:48:20.331005 ignition[1043]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:20.331005 ignition[1043]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:20.331005 ignition[1043]: INFO : mount: mount passed Mar 7 01:48:20.331005 ignition[1043]: INFO : Ignition finished successfully Mar 7 01:48:20.334920 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 7 01:48:20.360508 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 7 01:48:20.369242 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 7 01:48:20.383579 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 01:48:20.398454 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by mount (1056) Mar 7 01:48:20.408988 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:20.409020 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:20.412330 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:48:20.419456 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:48:20.421111 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 01:48:20.442477 ignition[1074]: INFO : Ignition 2.19.0 Mar 7 01:48:20.442477 ignition[1074]: INFO : Stage: files Mar 7 01:48:20.448655 ignition[1074]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:20.448655 ignition[1074]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:20.448655 ignition[1074]: DEBUG : files: compiled without relabeling support, skipping Mar 7 01:48:20.463114 ignition[1074]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 7 01:48:20.463114 ignition[1074]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 7 01:48:20.569379 ignition[1074]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Mar 7 01:48:20.569783 unknown[1074]: wrote ssh authorized keys file for user: core Mar 7 01:48:20.636608 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 7 01:48:20.773105 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 7 01:48:20.773105 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-arm64.raw: attempt #1 Mar 7 01:48:21.273447 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 7 01:48:21.568354 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:21.568354 ignition[1074]: INFO : files: op(c): [started] processing unit "containerd.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(c): [finished] processing unit "containerd.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Mar 7 01:48:21.697253 ignition[1074]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 7 01:48:21.697253 ignition[1074]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 7 01:48:21.697253 ignition[1074]: INFO : files: files passed Mar 7 01:48:21.697253 ignition[1074]: INFO : Ignition finished successfully Mar 7 01:48:21.619546 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 7 01:48:21.662320 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 7 01:48:21.673592 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 7 01:48:21.741899 initrd-setup-root-after-ignition[1100]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:48:21.741899 initrd-setup-root-after-ignition[1100]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:48:21.706673 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 7 01:48:21.770347 initrd-setup-root-after-ignition[1104]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:48:21.706769 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 7 01:48:21.738231 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 01:48:21.748989 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 7 01:48:21.781721 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 7 01:48:21.816060 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 7 01:48:21.816216 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 7 01:48:21.825569 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 7 01:48:21.835652 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 7 01:48:21.844634 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 7 01:48:21.861704 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 7 01:48:21.879370 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 01:48:21.894569 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 7 01:48:21.910817 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:48:21.916130 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:48:21.925964 systemd[1]: Stopped target timers.target - Timer Units. Mar 7 01:48:21.934334 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 7 01:48:21.934461 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 01:48:21.947093 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 7 01:48:21.951747 systemd[1]: Stopped target basic.target - Basic System. Mar 7 01:48:21.960561 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 7 01:48:21.969701 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 01:48:21.978105 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 7 01:48:21.987199 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 7 01:48:21.995968 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 01:48:22.005805 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 7 01:48:22.014129 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 7 01:48:22.023142 systemd[1]: Stopped target swap.target - Swaps. Mar 7 01:48:22.031039 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 7 01:48:22.031160 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 7 01:48:22.043753 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:48:22.049251 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:48:22.058815 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 7 01:48:22.058888 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:48:22.069569 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 7 01:48:22.069686 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 7 01:48:22.083162 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 7 01:48:22.083297 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 01:48:22.092812 systemd[1]: ignition-files.service: Deactivated successfully. Mar 7 01:48:22.092901 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 7 01:48:22.103135 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 7 01:48:22.152715 ignition[1125]: INFO : Ignition 2.19.0 Mar 7 01:48:22.152715 ignition[1125]: INFO : Stage: umount Mar 7 01:48:22.152715 ignition[1125]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:22.152715 ignition[1125]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:22.152715 ignition[1125]: INFO : umount: umount passed Mar 7 01:48:22.152715 ignition[1125]: INFO : Ignition finished successfully Mar 7 01:48:22.103219 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 01:48:22.126702 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 7 01:48:22.136326 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 7 01:48:22.140855 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:48:22.154681 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 7 01:48:22.161852 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 7 01:48:22.162001 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:48:22.169415 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 7 01:48:22.169540 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 01:48:22.186596 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 7 01:48:22.186699 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 7 01:48:22.201059 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 7 01:48:22.201168 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 7 01:48:22.208799 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 7 01:48:22.208864 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 7 01:48:22.222522 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 7 01:48:22.222590 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 7 01:48:22.231685 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 7 01:48:22.231738 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 7 01:48:22.236040 systemd[1]: Stopped target network.target - Network. Mar 7 01:48:22.244403 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 7 01:48:22.244468 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 01:48:22.257060 systemd[1]: Stopped target paths.target - Path Units. Mar 7 01:48:22.265497 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 7 01:48:22.269489 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:48:22.274889 systemd[1]: Stopped target slices.target - Slice Units. Mar 7 01:48:22.283144 systemd[1]: Stopped target sockets.target - Socket Units. Mar 7 01:48:22.291849 systemd[1]: iscsid.socket: Deactivated successfully. Mar 7 01:48:22.291907 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 01:48:22.301575 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 7 01:48:22.301625 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 01:48:22.310699 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 7 01:48:22.310759 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 7 01:48:22.315327 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 7 01:48:22.315363 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 7 01:48:22.323776 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 7 01:48:22.332866 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 7 01:48:22.336480 systemd-networkd[883]: eth0: DHCPv6 lease lost Mar 7 01:48:22.345950 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 7 01:48:22.346608 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 7 01:48:22.346724 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 7 01:48:22.542076 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: Data path switched from VF: enP17750s1 Mar 7 01:48:22.357558 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 7 01:48:22.359463 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 7 01:48:22.368078 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 7 01:48:22.368135 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:48:22.393641 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 7 01:48:22.400761 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 7 01:48:22.400838 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 01:48:22.410869 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 7 01:48:22.410915 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:48:22.421599 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 7 01:48:22.421637 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 7 01:48:22.429672 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 7 01:48:22.429707 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:48:22.440672 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:48:22.469846 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 7 01:48:22.470935 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:48:22.478812 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 7 01:48:22.478876 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 7 01:48:22.487225 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 7 01:48:22.487266 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:48:22.495563 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 7 01:48:22.495612 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 7 01:48:22.510275 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 7 01:48:22.510323 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 7 01:48:22.522068 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:48:22.522128 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:22.553680 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 7 01:48:22.562515 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 7 01:48:22.562588 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:48:22.574503 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:48:22.574550 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:22.584018 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 7 01:48:22.584120 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 7 01:48:22.600272 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 7 01:48:22.600451 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 7 01:48:22.654837 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 7 01:48:22.655149 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 7 01:48:22.663364 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 7 01:48:22.671662 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 7 01:48:22.671717 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 7 01:48:22.702714 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 7 01:48:22.996139 systemd[1]: Switching root. Mar 7 01:48:23.031026 systemd-journald[218]: Journal stopped Mar 7 01:48:11.171208 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Mar 7 01:48:11.171228 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Fri Mar 6 22:59:59 -00 2026 Mar 7 01:48:11.171236 kernel: KASLR enabled Mar 7 01:48:11.171242 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Mar 7 01:48:11.171249 kernel: printk: bootconsole [pl11] enabled Mar 7 01:48:11.171255 kernel: efi: EFI v2.7 by EDK II Mar 7 01:48:11.171262 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3f215018 RNG=0x3fd5f998 MEMRESERVE=0x3e44ee18 Mar 7 01:48:11.171268 kernel: random: crng init done Mar 7 01:48:11.171274 kernel: ACPI: Early table checksum verification disabled Mar 7 01:48:11.171280 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) Mar 7 01:48:11.171286 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171292 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171300 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) Mar 7 01:48:11.171306 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171313 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171320 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171326 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171334 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171340 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171347 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Mar 7 01:48:11.171353 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 7 01:48:11.171360 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Mar 7 01:48:11.171366 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Mar 7 01:48:11.171372 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] Mar 7 01:48:11.171379 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] Mar 7 01:48:11.171385 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] Mar 7 01:48:11.171392 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] Mar 7 01:48:11.171398 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] Mar 7 01:48:11.171406 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] Mar 7 01:48:11.171413 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] Mar 7 01:48:11.171419 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] Mar 7 01:48:11.171425 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] Mar 7 01:48:11.171432 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] Mar 7 01:48:11.171438 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] Mar 7 01:48:11.171444 kernel: NUMA: NODE_DATA [mem 0x1bf7ef800-0x1bf7f4fff] Mar 7 01:48:11.171451 kernel: Zone ranges: Mar 7 01:48:11.171457 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Mar 7 01:48:11.171463 kernel: DMA32 empty Mar 7 01:48:11.171470 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Mar 7 01:48:11.171476 kernel: Movable zone start for each node Mar 7 01:48:11.171487 kernel: Early memory node ranges Mar 7 01:48:11.171494 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Mar 7 01:48:11.171501 kernel: node 0: [mem 0x0000000000824000-0x000000003e54ffff] Mar 7 01:48:11.171507 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] Mar 7 01:48:11.171514 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] Mar 7 01:48:11.171522 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] Mar 7 01:48:11.171529 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] Mar 7 01:48:11.171536 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Mar 7 01:48:11.171543 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Mar 7 01:48:11.171550 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Mar 7 01:48:11.171556 kernel: psci: probing for conduit method from ACPI. Mar 7 01:48:11.171563 kernel: psci: PSCIv1.1 detected in firmware. Mar 7 01:48:11.171570 kernel: psci: Using standard PSCI v0.2 function IDs Mar 7 01:48:11.171577 kernel: psci: MIGRATE_INFO_TYPE not supported. Mar 7 01:48:11.171584 kernel: psci: SMC Calling Convention v1.4 Mar 7 01:48:11.171591 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Mar 7 01:48:11.171598 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Mar 7 01:48:11.171607 kernel: percpu: Embedded 30 pages/cpu s85736 r8192 d28952 u122880 Mar 7 01:48:11.171614 kernel: pcpu-alloc: s85736 r8192 d28952 u122880 alloc=30*4096 Mar 7 01:48:11.171620 kernel: pcpu-alloc: [0] 0 [0] 1 Mar 7 01:48:11.171627 kernel: Detected PIPT I-cache on CPU0 Mar 7 01:48:11.171634 kernel: CPU features: detected: GIC system register CPU interface Mar 7 01:48:11.171641 kernel: CPU features: detected: Hardware dirty bit management Mar 7 01:48:11.171648 kernel: CPU features: detected: Spectre-BHB Mar 7 01:48:11.171655 kernel: CPU features: kernel page table isolation forced ON by KASLR Mar 7 01:48:11.171662 kernel: CPU features: detected: Kernel page table isolation (KPTI) Mar 7 01:48:11.171668 kernel: CPU features: detected: ARM erratum 1418040 Mar 7 01:48:11.171675 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Mar 7 01:48:11.171683 kernel: CPU features: detected: SSBS not fully self-synchronizing Mar 7 01:48:11.171690 kernel: alternatives: applying boot alternatives Mar 7 01:48:11.171698 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=9d22c40559a0d209dc0fcc2dfdd5ddf9671e6da0cc59463f610ba522f01325a6 Mar 7 01:48:11.171705 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 7 01:48:11.171712 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 7 01:48:11.171719 kernel: Fallback order for Node 0: 0 Mar 7 01:48:11.171726 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Mar 7 01:48:11.171732 kernel: Policy zone: Normal Mar 7 01:48:11.171739 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 7 01:48:11.171746 kernel: software IO TLB: area num 2. Mar 7 01:48:11.171753 kernel: software IO TLB: mapped [mem 0x000000003a44e000-0x000000003e44e000] (64MB) Mar 7 01:48:11.171761 kernel: Memory: 3982636K/4194160K available (10304K kernel code, 2180K rwdata, 8116K rodata, 39424K init, 897K bss, 211524K reserved, 0K cma-reserved) Mar 7 01:48:11.171768 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 7 01:48:11.171775 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 7 01:48:11.171782 kernel: rcu: RCU event tracing is enabled. Mar 7 01:48:11.171790 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 7 01:48:11.171796 kernel: Trampoline variant of Tasks RCU enabled. Mar 7 01:48:11.171804 kernel: Tracing variant of Tasks RCU enabled. Mar 7 01:48:11.171811 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 7 01:48:11.171818 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 7 01:48:11.171825 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 7 01:48:11.171832 kernel: GICv3: 960 SPIs implemented Mar 7 01:48:11.171841 kernel: GICv3: 0 Extended SPIs implemented Mar 7 01:48:11.171848 kernel: Root IRQ handler: gic_handle_irq Mar 7 01:48:11.171854 kernel: GICv3: GICv3 features: 16 PPIs, RSS Mar 7 01:48:11.171861 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Mar 7 01:48:11.171868 kernel: ITS: No ITS available, not enabling LPIs Mar 7 01:48:11.171875 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 7 01:48:11.171882 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 7 01:48:11.171889 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Mar 7 01:48:11.171896 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Mar 7 01:48:11.171903 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Mar 7 01:48:11.171910 kernel: Console: colour dummy device 80x25 Mar 7 01:48:11.171918 kernel: printk: console [tty1] enabled Mar 7 01:48:11.171925 kernel: ACPI: Core revision 20230628 Mar 7 01:48:11.171933 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Mar 7 01:48:11.171940 kernel: pid_max: default: 32768 minimum: 301 Mar 7 01:48:11.171947 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 7 01:48:11.171954 kernel: landlock: Up and running. Mar 7 01:48:11.171961 kernel: SELinux: Initializing. Mar 7 01:48:11.171968 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.171975 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.171984 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 01:48:11.171991 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 01:48:11.171998 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0x100000e, misc 0x31e1 Mar 7 01:48:11.172005 kernel: Hyper-V: Host Build 10.0.26100.1480-1-0 Mar 7 01:48:11.172012 kernel: Hyper-V: enabling crash_kexec_post_notifiers Mar 7 01:48:11.172019 kernel: rcu: Hierarchical SRCU implementation. Mar 7 01:48:11.172026 kernel: rcu: Max phase no-delay instances is 400. Mar 7 01:48:11.172034 kernel: Remapping and enabling EFI services. Mar 7 01:48:11.172046 kernel: smp: Bringing up secondary CPUs ... Mar 7 01:48:11.172054 kernel: Detected PIPT I-cache on CPU1 Mar 7 01:48:11.172061 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Mar 7 01:48:11.172075 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 7 01:48:11.172085 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Mar 7 01:48:11.172092 kernel: smp: Brought up 1 node, 2 CPUs Mar 7 01:48:11.172100 kernel: SMP: Total of 2 processors activated. Mar 7 01:48:11.172108 kernel: CPU features: detected: 32-bit EL0 Support Mar 7 01:48:11.172115 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Mar 7 01:48:11.172124 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Mar 7 01:48:11.172131 kernel: CPU features: detected: CRC32 instructions Mar 7 01:48:11.172139 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Mar 7 01:48:11.172146 kernel: CPU features: detected: LSE atomic instructions Mar 7 01:48:11.172154 kernel: CPU features: detected: Privileged Access Never Mar 7 01:48:11.172161 kernel: CPU: All CPU(s) started at EL1 Mar 7 01:48:11.172168 kernel: alternatives: applying system-wide alternatives Mar 7 01:48:11.172175 kernel: devtmpfs: initialized Mar 7 01:48:11.172183 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 7 01:48:11.172192 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 7 01:48:11.172199 kernel: pinctrl core: initialized pinctrl subsystem Mar 7 01:48:11.172206 kernel: SMBIOS 3.1.0 present. Mar 7 01:48:11.172214 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 Mar 7 01:48:11.172221 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 7 01:48:11.172229 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 7 01:48:11.172236 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 7 01:48:11.172244 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 7 01:48:11.172251 kernel: audit: initializing netlink subsys (disabled) Mar 7 01:48:11.172260 kernel: audit: type=2000 audit(0.047:1): state=initialized audit_enabled=0 res=1 Mar 7 01:48:11.172268 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 7 01:48:11.172275 kernel: cpuidle: using governor menu Mar 7 01:48:11.172283 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 7 01:48:11.172291 kernel: ASID allocator initialised with 32768 entries Mar 7 01:48:11.172299 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 7 01:48:11.172306 kernel: Serial: AMBA PL011 UART driver Mar 7 01:48:11.172314 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Mar 7 01:48:11.172321 kernel: Modules: 0 pages in range for non-PLT usage Mar 7 01:48:11.172330 kernel: Modules: 509008 pages in range for PLT usage Mar 7 01:48:11.172338 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 7 01:48:11.172346 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Mar 7 01:48:11.172353 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Mar 7 01:48:11.172361 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Mar 7 01:48:11.172368 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 7 01:48:11.172375 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Mar 7 01:48:11.172383 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Mar 7 01:48:11.172390 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Mar 7 01:48:11.172399 kernel: ACPI: Added _OSI(Module Device) Mar 7 01:48:11.172406 kernel: ACPI: Added _OSI(Processor Device) Mar 7 01:48:11.172413 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 7 01:48:11.172420 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 7 01:48:11.172428 kernel: ACPI: Interpreter enabled Mar 7 01:48:11.172435 kernel: ACPI: Using GIC for interrupt routing Mar 7 01:48:11.172443 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Mar 7 01:48:11.172450 kernel: printk: console [ttyAMA0] enabled Mar 7 01:48:11.172457 kernel: printk: bootconsole [pl11] disabled Mar 7 01:48:11.172466 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Mar 7 01:48:11.172474 kernel: iommu: Default domain type: Translated Mar 7 01:48:11.172481 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 7 01:48:11.172488 kernel: efivars: Registered efivars operations Mar 7 01:48:11.172495 kernel: vgaarb: loaded Mar 7 01:48:11.172503 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 7 01:48:11.172510 kernel: VFS: Disk quotas dquot_6.6.0 Mar 7 01:48:11.172517 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 7 01:48:11.172525 kernel: pnp: PnP ACPI init Mar 7 01:48:11.172533 kernel: pnp: PnP ACPI: found 0 devices Mar 7 01:48:11.172540 kernel: NET: Registered PF_INET protocol family Mar 7 01:48:11.172548 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 7 01:48:11.172555 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 7 01:48:11.172563 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 7 01:48:11.172570 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 7 01:48:11.172577 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 7 01:48:11.172585 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 7 01:48:11.172592 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.172601 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 01:48:11.172608 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 7 01:48:11.172616 kernel: PCI: CLS 0 bytes, default 64 Mar 7 01:48:11.172623 kernel: kvm [1]: HYP mode not available Mar 7 01:48:11.172630 kernel: Initialise system trusted keyrings Mar 7 01:48:11.172638 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 7 01:48:11.172645 kernel: Key type asymmetric registered Mar 7 01:48:11.172652 kernel: Asymmetric key parser 'x509' registered Mar 7 01:48:11.172659 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 7 01:48:11.172668 kernel: io scheduler mq-deadline registered Mar 7 01:48:11.172676 kernel: io scheduler kyber registered Mar 7 01:48:11.172683 kernel: io scheduler bfq registered Mar 7 01:48:11.172690 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 7 01:48:11.172697 kernel: thunder_xcv, ver 1.0 Mar 7 01:48:11.172705 kernel: thunder_bgx, ver 1.0 Mar 7 01:48:11.172712 kernel: nicpf, ver 1.0 Mar 7 01:48:11.172719 kernel: nicvf, ver 1.0 Mar 7 01:48:11.172837 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 7 01:48:11.172916 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-03-07T01:48:10 UTC (1772848090) Mar 7 01:48:11.172926 kernel: efifb: probing for efifb Mar 7 01:48:11.172934 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Mar 7 01:48:11.172941 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Mar 7 01:48:11.172949 kernel: efifb: scrolling: redraw Mar 7 01:48:11.172956 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 7 01:48:11.172963 kernel: Console: switching to colour frame buffer device 128x48 Mar 7 01:48:11.172971 kernel: fb0: EFI VGA frame buffer device Mar 7 01:48:11.172980 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Mar 7 01:48:11.172989 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 7 01:48:11.172996 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 6 counters available Mar 7 01:48:11.173004 kernel: watchdog: Delayed init of the lockup detector failed: -19 Mar 7 01:48:11.173011 kernel: watchdog: Hard watchdog permanently disabled Mar 7 01:48:11.173018 kernel: NET: Registered PF_INET6 protocol family Mar 7 01:48:11.173025 kernel: Segment Routing with IPv6 Mar 7 01:48:11.173033 kernel: In-situ OAM (IOAM) with IPv6 Mar 7 01:48:11.173040 kernel: NET: Registered PF_PACKET protocol family Mar 7 01:48:11.173049 kernel: Key type dns_resolver registered Mar 7 01:48:11.173056 kernel: registered taskstats version 1 Mar 7 01:48:11.173068 kernel: Loading compiled-in X.509 certificates Mar 7 01:48:11.173077 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: e62b4e4ebcb406beff1271ecc7444548c4ab67e9' Mar 7 01:48:11.173084 kernel: Key type .fscrypt registered Mar 7 01:48:11.173091 kernel: Key type fscrypt-provisioning registered Mar 7 01:48:11.173098 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 7 01:48:11.173106 kernel: ima: Allocated hash algorithm: sha1 Mar 7 01:48:11.173113 kernel: ima: No architecture policies found Mar 7 01:48:11.173122 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 7 01:48:11.173130 kernel: clk: Disabling unused clocks Mar 7 01:48:11.173137 kernel: Freeing unused kernel memory: 39424K Mar 7 01:48:11.173144 kernel: Run /init as init process Mar 7 01:48:11.173151 kernel: with arguments: Mar 7 01:48:11.173158 kernel: /init Mar 7 01:48:11.173165 kernel: with environment: Mar 7 01:48:11.173173 kernel: HOME=/ Mar 7 01:48:11.173180 kernel: TERM=linux Mar 7 01:48:11.173189 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 7 01:48:11.173200 systemd[1]: Detected virtualization microsoft. Mar 7 01:48:11.173208 systemd[1]: Detected architecture arm64. Mar 7 01:48:11.173216 systemd[1]: Running in initrd. Mar 7 01:48:11.173224 systemd[1]: No hostname configured, using default hostname. Mar 7 01:48:11.173231 systemd[1]: Hostname set to . Mar 7 01:48:11.173239 systemd[1]: Initializing machine ID from random generator. Mar 7 01:48:11.173247 systemd[1]: Queued start job for default target initrd.target. Mar 7 01:48:11.173257 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:48:11.173265 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:48:11.173274 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 7 01:48:11.173282 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 7 01:48:11.173290 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 7 01:48:11.173298 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 7 01:48:11.173307 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 7 01:48:11.173317 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 7 01:48:11.173325 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:48:11.173333 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:48:11.173343 systemd[1]: Reached target paths.target - Path Units. Mar 7 01:48:11.173351 systemd[1]: Reached target slices.target - Slice Units. Mar 7 01:48:11.173359 systemd[1]: Reached target swap.target - Swaps. Mar 7 01:48:11.173367 systemd[1]: Reached target timers.target - Timer Units. Mar 7 01:48:11.173375 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 01:48:11.173385 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 01:48:11.173393 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 7 01:48:11.173401 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 7 01:48:11.173409 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:48:11.173417 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 7 01:48:11.173424 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:48:11.173432 systemd[1]: Reached target sockets.target - Socket Units. Mar 7 01:48:11.173440 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 7 01:48:11.173450 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 7 01:48:11.173457 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 7 01:48:11.173465 systemd[1]: Starting systemd-fsck-usr.service... Mar 7 01:48:11.173473 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 7 01:48:11.173496 systemd-journald[218]: Collecting audit messages is disabled. Mar 7 01:48:11.173516 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 7 01:48:11.173524 systemd-journald[218]: Journal started Mar 7 01:48:11.173542 systemd-journald[218]: Runtime Journal (/run/log/journal/9be137b8f6ae4d63975e41ca0ff2e618) is 8.0M, max 78.5M, 70.5M free. Mar 7 01:48:11.184997 systemd-modules-load[219]: Inserted module 'overlay' Mar 7 01:48:11.189739 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:11.204468 systemd[1]: Started systemd-journald.service - Journal Service. Mar 7 01:48:11.205434 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 7 01:48:11.227157 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 7 01:48:11.227186 kernel: Bridge firewalling registered Mar 7 01:48:11.218048 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:48:11.226140 systemd-modules-load[219]: Inserted module 'br_netfilter' Mar 7 01:48:11.232084 systemd[1]: Finished systemd-fsck-usr.service. Mar 7 01:48:11.239099 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 7 01:48:11.247278 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:11.269323 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:11.281669 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 01:48:11.297288 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 7 01:48:11.313243 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 7 01:48:11.320860 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:11.333585 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:48:11.350161 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 7 01:48:11.356119 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:48:11.375267 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 7 01:48:11.381237 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 7 01:48:11.403881 dracut-cmdline[253]: dracut-dracut-053 Mar 7 01:48:11.405207 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 7 01:48:11.425391 dracut-cmdline[253]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=9d22c40559a0d209dc0fcc2dfdd5ddf9671e6da0cc59463f610ba522f01325a6 Mar 7 01:48:11.417609 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:48:11.430952 systemd-resolved[254]: Positive Trust Anchors: Mar 7 01:48:11.430962 systemd-resolved[254]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 7 01:48:11.430994 systemd-resolved[254]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 7 01:48:11.433865 systemd-resolved[254]: Defaulting to hostname 'linux'. Mar 7 01:48:11.461475 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 7 01:48:11.469530 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:48:11.574089 kernel: SCSI subsystem initialized Mar 7 01:48:11.583079 kernel: Loading iSCSI transport class v2.0-870. Mar 7 01:48:11.591094 kernel: iscsi: registered transport (tcp) Mar 7 01:48:11.606888 kernel: iscsi: registered transport (qla4xxx) Mar 7 01:48:11.606929 kernel: QLogic iSCSI HBA Driver Mar 7 01:48:11.645115 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 7 01:48:11.655362 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 7 01:48:11.683865 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 7 01:48:11.683923 kernel: device-mapper: uevent: version 1.0.3 Mar 7 01:48:11.689123 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 7 01:48:11.738093 kernel: raid6: neonx8 gen() 15785 MB/s Mar 7 01:48:11.755080 kernel: raid6: neonx4 gen() 15694 MB/s Mar 7 01:48:11.774072 kernel: raid6: neonx2 gen() 13277 MB/s Mar 7 01:48:11.796089 kernel: raid6: neonx1 gen() 10538 MB/s Mar 7 01:48:11.811076 kernel: raid6: int64x8 gen() 6580 MB/s Mar 7 01:48:11.830072 kernel: raid6: int64x4 gen() 7371 MB/s Mar 7 01:48:11.850074 kernel: raid6: int64x2 gen() 6146 MB/s Mar 7 01:48:11.872674 kernel: raid6: int64x1 gen() 5071 MB/s Mar 7 01:48:11.872695 kernel: raid6: using algorithm neonx8 gen() 15785 MB/s Mar 7 01:48:11.894440 kernel: raid6: .... xor() 12041 MB/s, rmw enabled Mar 7 01:48:11.894460 kernel: raid6: using neon recovery algorithm Mar 7 01:48:11.905191 kernel: xor: measuring software checksum speed Mar 7 01:48:11.905221 kernel: 8regs : 19797 MB/sec Mar 7 01:48:11.907912 kernel: 32regs : 19688 MB/sec Mar 7 01:48:11.910621 kernel: arm64_neon : 27105 MB/sec Mar 7 01:48:11.913795 kernel: xor: using function: arm64_neon (27105 MB/sec) Mar 7 01:48:11.964091 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 7 01:48:11.974452 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 7 01:48:11.989216 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:48:12.009302 systemd-udevd[439]: Using default interface naming scheme 'v255'. Mar 7 01:48:12.013731 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:48:12.030212 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 7 01:48:12.044365 dracut-pre-trigger[452]: rd.md=0: removing MD RAID activation Mar 7 01:48:12.071625 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 01:48:12.083515 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 7 01:48:12.120835 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:48:12.136221 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 7 01:48:12.157159 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 7 01:48:12.172348 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 01:48:12.180673 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:48:12.196195 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 7 01:48:12.219647 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 7 01:48:12.237303 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:48:12.247317 kernel: hv_vmbus: Vmbus version:5.3 Mar 7 01:48:12.237413 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:12.275792 kernel: hv_vmbus: registering driver hyperv_keyboard Mar 7 01:48:12.275814 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Mar 7 01:48:12.275492 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:12.291114 kernel: pps_core: LinuxPPS API ver. 1 registered Mar 7 01:48:12.282405 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:48:12.310418 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Mar 7 01:48:12.310441 kernel: PTP clock support registered Mar 7 01:48:12.282564 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.303334 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.332382 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.360718 kernel: hv_vmbus: registering driver hid_hyperv Mar 7 01:48:12.360743 kernel: hv_utils: Registering HyperV Utility Driver Mar 7 01:48:12.360753 kernel: hv_vmbus: registering driver hv_utils Mar 7 01:48:12.360763 kernel: hv_vmbus: registering driver hv_storvsc Mar 7 01:48:12.360773 kernel: hv_utils: Heartbeat IC version 3.0 Mar 7 01:48:12.360782 kernel: hv_utils: Shutdown IC version 3.2 Mar 7 01:48:12.340458 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 7 01:48:11.970378 kernel: hv_utils: TimeSync IC version 4.0 Mar 7 01:48:11.981887 kernel: hv_vmbus: registering driver hv_netvsc Mar 7 01:48:11.981904 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Mar 7 01:48:11.981913 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Mar 7 01:48:11.982028 systemd-journald[218]: Time jumped backwards, rotating. Mar 7 01:48:11.958860 systemd-resolved[254]: Clock change detected. Flushing caches. Mar 7 01:48:12.003302 kernel: scsi host0: storvsc_host_t Mar 7 01:48:12.009316 kernel: scsi host1: storvsc_host_t Mar 7 01:48:12.009507 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Mar 7 01:48:12.009543 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 5 Mar 7 01:48:11.960883 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.003912 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:12.027171 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:48:12.027305 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:12.039856 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:48:12.072201 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Mar 7 01:48:12.072388 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: VF slot 1 added Mar 7 01:48:12.072502 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 7 01:48:12.039920 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.047003 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.099649 kernel: hv_vmbus: registering driver hv_pci Mar 7 01:48:12.099671 kernel: hv_pci d7ed302e-4556-46d5-ba91-8abdbb0cefa4: PCI VMBus probing: Using version 0x10004 Mar 7 01:48:12.079703 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:12.110110 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Mar 7 01:48:12.110300 kernel: hv_pci d7ed302e-4556-46d5-ba91-8abdbb0cefa4: PCI host bridge to bus 4556:00 Mar 7 01:48:12.118410 kernel: pci_bus 4556:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Mar 7 01:48:12.129181 kernel: pci_bus 4556:00: No busn resource found for root bus, will use [bus 00-ff] Mar 7 01:48:12.129359 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Mar 7 01:48:12.129487 kernel: pci 4556:00:02.0: [15b3:1018] type 00 class 0x020000 Mar 7 01:48:12.137150 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Mar 7 01:48:12.137660 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:12.163031 kernel: sd 0:0:0:0: [sda] Write Protect is off Mar 7 01:48:12.163199 kernel: pci 4556:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Mar 7 01:48:12.163220 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Mar 7 01:48:12.163309 kernel: pci 4556:00:02.0: enabling Extended Tags Mar 7 01:48:12.163323 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Mar 7 01:48:12.165669 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 01:48:12.213143 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 01:48:12.213166 kernel: pci 4556:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 4556:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Mar 7 01:48:12.213346 kernel: pci_bus 4556:00: busn_res: [bus 00-ff] end is updated to 00 Mar 7 01:48:12.213466 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#218 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 7 01:48:12.213574 kernel: pci 4556:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Mar 7 01:48:12.213666 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Mar 7 01:48:12.224324 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:12.258894 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#255 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 7 01:48:12.268353 kernel: mlx5_core 4556:00:02.0: enabling device (0000 -> 0002) Mar 7 01:48:12.274513 kernel: mlx5_core 4556:00:02.0: firmware version: 16.30.5026 Mar 7 01:48:12.475243 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: VF registering: eth1 Mar 7 01:48:12.475431 kernel: mlx5_core 4556:00:02.0 eth1: joined to eth0 Mar 7 01:48:12.480505 kernel: mlx5_core 4556:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Mar 7 01:48:12.490466 kernel: mlx5_core 4556:00:02.0 enP17750s1: renamed from eth1 Mar 7 01:48:12.901461 kernel: BTRFS: device fsid 237c8587-8110-47ef-99f9-37e4ed4d3b31 devid 1 transid 36 /dev/sda3 scanned by (udev-worker) (488) Mar 7 01:48:12.914815 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Mar 7 01:48:12.920072 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Mar 7 01:48:12.952618 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 7 01:48:13.969912 disk-uuid[597]: Warning: The kernel is still using the old partition table. Mar 7 01:48:13.969912 disk-uuid[597]: The new table will be used at the next reboot or after you Mar 7 01:48:13.969912 disk-uuid[597]: run partprobe(8) or kpartx(8) Mar 7 01:48:13.969912 disk-uuid[597]: The operation has completed successfully. Mar 7 01:48:14.848642 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by (udev-worker) (512) Mar 7 01:48:14.863055 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Mar 7 01:48:15.039137 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Mar 7 01:48:15.358722 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Mar 7 01:48:15.828505 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 7 01:48:15.828614 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 7 01:48:15.846574 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 7 01:48:15.856478 sh[699]: Success Mar 7 01:48:15.887510 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Mar 7 01:48:16.178336 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 7 01:48:16.183016 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 7 01:48:16.195572 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 7 01:48:16.230009 kernel: BTRFS info (device dm-0): first mount of filesystem 237c8587-8110-47ef-99f9-37e4ed4d3b31 Mar 7 01:48:16.230061 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:16.235380 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 7 01:48:16.239375 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 7 01:48:16.242657 kernel: BTRFS info (device dm-0): using free space tree Mar 7 01:48:16.582941 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 7 01:48:16.587212 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 7 01:48:16.608748 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 7 01:48:16.614589 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 7 01:48:16.648644 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:16.648698 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:16.652118 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:48:16.693472 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:48:16.702291 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 7 01:48:16.711938 kernel: BTRFS info (device sda6): last unmount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:16.719250 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 7 01:48:16.732684 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 7 01:48:16.754327 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 01:48:16.769558 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 7 01:48:16.793733 systemd-networkd[883]: lo: Link UP Mar 7 01:48:16.793742 systemd-networkd[883]: lo: Gained carrier Mar 7 01:48:16.795386 systemd-networkd[883]: Enumeration completed Mar 7 01:48:16.795923 systemd-networkd[883]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:48:16.795926 systemd-networkd[883]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:48:16.800092 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 7 01:48:16.805381 systemd[1]: Reached target network.target - Network. Mar 7 01:48:16.876540 kernel: mlx5_core 4556:00:02.0 enP17750s1: Link up Mar 7 01:48:16.916719 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: Data path switched to VF: enP17750s1 Mar 7 01:48:16.917240 systemd-networkd[883]: enP17750s1: Link UP Mar 7 01:48:16.917340 systemd-networkd[883]: eth0: Link UP Mar 7 01:48:16.917503 systemd-networkd[883]: eth0: Gained carrier Mar 7 01:48:16.917512 systemd-networkd[883]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:48:16.926615 systemd-networkd[883]: enP17750s1: Gained carrier Mar 7 01:48:16.945488 systemd-networkd[883]: eth0: DHCPv4 address 10.200.20.15/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 7 01:48:17.565917 ignition[863]: Ignition 2.19.0 Mar 7 01:48:17.565928 ignition[863]: Stage: fetch-offline Mar 7 01:48:17.570058 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 01:48:17.565964 ignition[863]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.565972 ignition[863]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.566062 ignition[863]: parsed url from cmdline: "" Mar 7 01:48:17.566064 ignition[863]: no config URL provided Mar 7 01:48:17.566069 ignition[863]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 01:48:17.591708 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 7 01:48:17.566076 ignition[863]: no config at "/usr/lib/ignition/user.ign" Mar 7 01:48:17.566080 ignition[863]: failed to fetch config: resource requires networking Mar 7 01:48:17.566410 ignition[863]: Ignition finished successfully Mar 7 01:48:17.610690 ignition[893]: Ignition 2.19.0 Mar 7 01:48:17.610697 ignition[893]: Stage: fetch Mar 7 01:48:17.610889 ignition[893]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.610898 ignition[893]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.610999 ignition[893]: parsed url from cmdline: "" Mar 7 01:48:17.611002 ignition[893]: no config URL provided Mar 7 01:48:17.611006 ignition[893]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 01:48:17.611014 ignition[893]: no config at "/usr/lib/ignition/user.ign" Mar 7 01:48:17.611036 ignition[893]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Mar 7 01:48:17.697435 ignition[893]: GET result: OK Mar 7 01:48:17.697532 ignition[893]: config has been read from IMDS userdata Mar 7 01:48:17.697579 ignition[893]: parsing config with SHA512: 91949a450c5096514d3c2c8c0d84e48291905018ba2b69f9b8c000712f659a4ad5d8dfc05d829a160b79aff16badf2df05c541342e2d09b31f0c75a1fc715f3c Mar 7 01:48:17.701744 unknown[893]: fetched base config from "system" Mar 7 01:48:17.702160 ignition[893]: fetch: fetch complete Mar 7 01:48:17.701751 unknown[893]: fetched base config from "system" Mar 7 01:48:17.702165 ignition[893]: fetch: fetch passed Mar 7 01:48:17.701756 unknown[893]: fetched user config from "azure" Mar 7 01:48:17.702204 ignition[893]: Ignition finished successfully Mar 7 01:48:17.705492 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 7 01:48:17.726629 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 7 01:48:17.744131 ignition[899]: Ignition 2.19.0 Mar 7 01:48:17.744141 ignition[899]: Stage: kargs Mar 7 01:48:17.744316 ignition[899]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.749608 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 7 01:48:17.744328 ignition[899]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.745409 ignition[899]: kargs: kargs passed Mar 7 01:48:17.745468 ignition[899]: Ignition finished successfully Mar 7 01:48:17.773000 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 7 01:48:17.788280 ignition[906]: Ignition 2.19.0 Mar 7 01:48:17.788289 ignition[906]: Stage: disks Mar 7 01:48:17.792096 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 7 01:48:17.788463 ignition[906]: no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:17.798663 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 7 01:48:17.788474 ignition[906]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:17.806764 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 7 01:48:17.789375 ignition[906]: disks: disks passed Mar 7 01:48:17.815170 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 7 01:48:17.789422 ignition[906]: Ignition finished successfully Mar 7 01:48:17.823596 systemd[1]: Reached target sysinit.target - System Initialization. Mar 7 01:48:17.832140 systemd[1]: Reached target basic.target - Basic System. Mar 7 01:48:17.850728 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 7 01:48:17.927014 systemd-fsck[914]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Mar 7 01:48:17.934709 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 7 01:48:17.948625 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 7 01:48:18.003455 kernel: EXT4-fs (sda9): mounted filesystem 596a8ea8-9d3d-4d06-a56e-9d3ebd3cb76d r/w with ordered data mode. Quota mode: none. Mar 7 01:48:18.004470 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 7 01:48:18.008032 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 7 01:48:18.060513 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 01:48:18.078459 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (925) Mar 7 01:48:18.089303 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:18.089357 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:18.092603 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:48:18.099455 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:48:18.100564 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 7 01:48:18.106615 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 7 01:48:18.113697 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 7 01:48:18.113731 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 01:48:18.124861 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 01:48:18.138269 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 7 01:48:18.154773 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 7 01:48:18.675477 coreos-metadata[942]: Mar 07 01:48:18.675 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 7 01:48:18.682708 coreos-metadata[942]: Mar 07 01:48:18.682 INFO Fetch successful Mar 7 01:48:18.686800 coreos-metadata[942]: Mar 07 01:48:18.686 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Mar 7 01:48:18.695240 systemd-networkd[883]: eth0: Gained IPv6LL Mar 7 01:48:18.705164 coreos-metadata[942]: Mar 07 01:48:18.705 INFO Fetch successful Mar 7 01:48:18.721984 coreos-metadata[942]: Mar 07 01:48:18.720 INFO wrote hostname ci-4081.3.6-n-e8b1fc37af to /sysroot/etc/hostname Mar 7 01:48:18.729407 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 01:48:18.962338 initrd-setup-root[955]: cut: /sysroot/etc/passwd: No such file or directory Mar 7 01:48:19.017719 initrd-setup-root[962]: cut: /sysroot/etc/group: No such file or directory Mar 7 01:48:19.039450 initrd-setup-root[969]: cut: /sysroot/etc/shadow: No such file or directory Mar 7 01:48:19.044813 initrd-setup-root[976]: cut: /sysroot/etc/gshadow: No such file or directory Mar 7 01:48:20.269426 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 7 01:48:20.281623 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 7 01:48:20.288583 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 7 01:48:20.305802 kernel: BTRFS info (device sda6): last unmount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:20.306332 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 7 01:48:20.327671 ignition[1043]: INFO : Ignition 2.19.0 Mar 7 01:48:20.331005 ignition[1043]: INFO : Stage: mount Mar 7 01:48:20.331005 ignition[1043]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:20.331005 ignition[1043]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:20.331005 ignition[1043]: INFO : mount: mount passed Mar 7 01:48:20.331005 ignition[1043]: INFO : Ignition finished successfully Mar 7 01:48:20.334920 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 7 01:48:20.360508 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 7 01:48:20.369242 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 7 01:48:20.383579 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 01:48:20.398454 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by mount (1056) Mar 7 01:48:20.408988 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 01:48:20.409020 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 01:48:20.412330 kernel: BTRFS info (device sda6): using free space tree Mar 7 01:48:20.419456 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 01:48:20.421111 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 01:48:20.442477 ignition[1074]: INFO : Ignition 2.19.0 Mar 7 01:48:20.442477 ignition[1074]: INFO : Stage: files Mar 7 01:48:20.448655 ignition[1074]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:20.448655 ignition[1074]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:20.448655 ignition[1074]: DEBUG : files: compiled without relabeling support, skipping Mar 7 01:48:20.463114 ignition[1074]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 7 01:48:20.463114 ignition[1074]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 7 01:48:20.569379 ignition[1074]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 7 01:48:20.575421 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Mar 7 01:48:20.569783 unknown[1074]: wrote ssh authorized keys file for user: core Mar 7 01:48:20.636608 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 7 01:48:20.773105 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 7 01:48:20.773105 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:20.788744 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-arm64.raw: attempt #1 Mar 7 01:48:21.273447 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 7 01:48:21.568354 ignition[1074]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 7 01:48:21.568354 ignition[1074]: INFO : files: op(c): [started] processing unit "containerd.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(c): [finished] processing unit "containerd.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Mar 7 01:48:21.616211 ignition[1074]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Mar 7 01:48:21.697253 ignition[1074]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 7 01:48:21.697253 ignition[1074]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 7 01:48:21.697253 ignition[1074]: INFO : files: files passed Mar 7 01:48:21.697253 ignition[1074]: INFO : Ignition finished successfully Mar 7 01:48:21.619546 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 7 01:48:21.662320 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 7 01:48:21.673592 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 7 01:48:21.741899 initrd-setup-root-after-ignition[1100]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:48:21.741899 initrd-setup-root-after-ignition[1100]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:48:21.706673 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 7 01:48:21.770347 initrd-setup-root-after-ignition[1104]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 01:48:21.706769 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 7 01:48:21.738231 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 01:48:21.748989 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 7 01:48:21.781721 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 7 01:48:21.816060 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 7 01:48:21.816216 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 7 01:48:21.825569 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 7 01:48:21.835652 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 7 01:48:21.844634 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 7 01:48:21.861704 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 7 01:48:21.879370 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 01:48:21.894569 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 7 01:48:21.910817 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:48:21.916130 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:48:21.925964 systemd[1]: Stopped target timers.target - Timer Units. Mar 7 01:48:21.934334 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 7 01:48:21.934461 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 01:48:21.947093 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 7 01:48:21.951747 systemd[1]: Stopped target basic.target - Basic System. Mar 7 01:48:21.960561 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 7 01:48:21.969701 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 01:48:21.978105 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 7 01:48:21.987199 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 7 01:48:21.995968 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 01:48:22.005805 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 7 01:48:22.014129 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 7 01:48:22.023142 systemd[1]: Stopped target swap.target - Swaps. Mar 7 01:48:22.031039 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 7 01:48:22.031160 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 7 01:48:22.043753 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:48:22.049251 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:48:22.058815 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 7 01:48:22.058888 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:48:22.069569 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 7 01:48:22.069686 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 7 01:48:22.083162 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 7 01:48:22.083297 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 01:48:22.092812 systemd[1]: ignition-files.service: Deactivated successfully. Mar 7 01:48:22.092901 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 7 01:48:22.103135 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 7 01:48:22.152715 ignition[1125]: INFO : Ignition 2.19.0 Mar 7 01:48:22.152715 ignition[1125]: INFO : Stage: umount Mar 7 01:48:22.152715 ignition[1125]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 01:48:22.152715 ignition[1125]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 7 01:48:22.152715 ignition[1125]: INFO : umount: umount passed Mar 7 01:48:22.152715 ignition[1125]: INFO : Ignition finished successfully Mar 7 01:48:22.103219 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 01:48:22.126702 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 7 01:48:22.136326 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 7 01:48:22.140855 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:48:22.154681 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 7 01:48:22.161852 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 7 01:48:22.162001 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:48:22.169415 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 7 01:48:22.169540 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 01:48:22.186596 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 7 01:48:22.186699 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 7 01:48:22.201059 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 7 01:48:22.201168 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 7 01:48:22.208799 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 7 01:48:22.208864 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 7 01:48:22.222522 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 7 01:48:22.222590 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 7 01:48:22.231685 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 7 01:48:22.231738 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 7 01:48:22.236040 systemd[1]: Stopped target network.target - Network. Mar 7 01:48:22.244403 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 7 01:48:22.244468 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 01:48:22.257060 systemd[1]: Stopped target paths.target - Path Units. Mar 7 01:48:22.265497 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 7 01:48:22.269489 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:48:22.274889 systemd[1]: Stopped target slices.target - Slice Units. Mar 7 01:48:22.283144 systemd[1]: Stopped target sockets.target - Socket Units. Mar 7 01:48:22.291849 systemd[1]: iscsid.socket: Deactivated successfully. Mar 7 01:48:22.291907 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 01:48:22.301575 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 7 01:48:22.301625 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 01:48:22.310699 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 7 01:48:22.310759 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 7 01:48:22.315327 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 7 01:48:22.315363 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 7 01:48:22.323776 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 7 01:48:22.332866 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 7 01:48:22.336480 systemd-networkd[883]: eth0: DHCPv6 lease lost Mar 7 01:48:22.345950 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 7 01:48:22.346608 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 7 01:48:22.346724 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 7 01:48:22.542076 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: Data path switched from VF: enP17750s1 Mar 7 01:48:22.357558 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 7 01:48:22.359463 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 7 01:48:22.368078 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 7 01:48:22.368135 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:48:22.393641 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 7 01:48:22.400761 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 7 01:48:22.400838 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 01:48:22.410869 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 7 01:48:22.410915 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:48:22.421599 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 7 01:48:22.421637 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 7 01:48:22.429672 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 7 01:48:22.429707 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:48:22.440672 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:48:22.469846 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 7 01:48:22.470935 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:48:22.478812 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 7 01:48:22.478876 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 7 01:48:22.487225 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 7 01:48:22.487266 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:48:22.495563 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 7 01:48:22.495612 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 7 01:48:22.510275 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 7 01:48:22.510323 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 7 01:48:22.522068 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 01:48:22.522128 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 01:48:22.553680 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 7 01:48:22.562515 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 7 01:48:22.562588 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:48:22.574503 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:48:22.574550 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:22.584018 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 7 01:48:22.584120 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 7 01:48:22.600272 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 7 01:48:22.600451 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 7 01:48:22.654837 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 7 01:48:22.655149 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 7 01:48:22.663364 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 7 01:48:22.671662 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 7 01:48:22.671717 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 7 01:48:22.702714 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 7 01:48:22.996139 systemd[1]: Switching root. Mar 7 01:48:23.031026 systemd-journald[218]: Journal stopped Mar 7 01:48:29.022224 systemd-journald[218]: Received SIGTERM from PID 1 (systemd). Mar 7 01:48:29.022260 kernel: SELinux: policy capability network_peer_controls=1 Mar 7 01:48:29.022271 kernel: SELinux: policy capability open_perms=1 Mar 7 01:48:29.022284 kernel: SELinux: policy capability extended_socket_class=1 Mar 7 01:48:29.022292 kernel: SELinux: policy capability always_check_network=0 Mar 7 01:48:29.022299 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 7 01:48:29.022308 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 7 01:48:29.022316 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 7 01:48:29.022324 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 7 01:48:29.022332 kernel: audit: type=1403 audit(1772848104.783:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 7 01:48:29.022343 systemd[1]: Successfully loaded SELinux policy in 138.458ms. Mar 7 01:48:29.022352 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.001ms. Mar 7 01:48:29.022362 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 7 01:48:29.022371 systemd[1]: Detected virtualization microsoft. Mar 7 01:48:29.022381 systemd[1]: Detected architecture arm64. Mar 7 01:48:29.022391 systemd[1]: Detected first boot. Mar 7 01:48:29.022400 systemd[1]: Hostname set to . Mar 7 01:48:29.022411 systemd[1]: Initializing machine ID from random generator. Mar 7 01:48:29.022420 zram_generator::config[1184]: No configuration found. Mar 7 01:48:29.022430 systemd[1]: Populated /etc with preset unit settings. Mar 7 01:48:29.022451 systemd[1]: Queued start job for default target multi-user.target. Mar 7 01:48:29.022465 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Mar 7 01:48:29.022475 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 7 01:48:29.022485 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 7 01:48:29.022494 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 7 01:48:29.022504 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 7 01:48:29.022513 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 7 01:48:29.022522 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 7 01:48:29.022533 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 7 01:48:29.022542 systemd[1]: Created slice user.slice - User and Session Slice. Mar 7 01:48:29.022552 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 01:48:29.022561 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 01:48:29.022570 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 7 01:48:29.022580 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 7 01:48:29.022589 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 7 01:48:29.022598 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 7 01:48:29.022608 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Mar 7 01:48:29.022619 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 01:48:29.022629 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 7 01:48:29.022638 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 01:48:29.022650 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 7 01:48:29.022660 systemd[1]: Reached target slices.target - Slice Units. Mar 7 01:48:29.022670 systemd[1]: Reached target swap.target - Swaps. Mar 7 01:48:29.022679 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 7 01:48:29.022690 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 7 01:48:29.022699 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 7 01:48:29.022709 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 7 01:48:29.022718 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 7 01:48:29.022727 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 7 01:48:29.022737 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 01:48:29.022746 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 7 01:48:29.022758 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 7 01:48:29.022767 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 7 01:48:29.022777 systemd[1]: Mounting media.mount - External Media Directory... Mar 7 01:48:29.022786 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 7 01:48:29.022795 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 7 01:48:29.022805 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 7 01:48:29.022816 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 7 01:48:29.022826 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:48:29.022837 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 7 01:48:29.022846 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 7 01:48:29.022855 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:48:29.022865 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 7 01:48:29.022874 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:48:29.022884 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 7 01:48:29.022893 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:48:29.022905 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 7 01:48:29.022915 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Mar 7 01:48:29.022925 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Mar 7 01:48:29.022934 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 7 01:48:29.022944 kernel: fuse: init (API version 7.39) Mar 7 01:48:29.022952 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 7 01:48:29.022961 kernel: loop: module loaded Mar 7 01:48:29.022970 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 7 01:48:29.022981 kernel: ACPI: bus type drm_connector registered Mar 7 01:48:29.022990 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 7 01:48:29.023027 systemd-journald[1303]: Collecting audit messages is disabled. Mar 7 01:48:29.023047 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 7 01:48:29.023059 systemd-journald[1303]: Journal started Mar 7 01:48:29.023079 systemd-journald[1303]: Runtime Journal (/run/log/journal/f284eccc0e2c463db1f6ea526d021a7c) is 8.0M, max 78.5M, 70.5M free. Mar 7 01:48:29.041210 systemd[1]: Started systemd-journald.service - Journal Service. Mar 7 01:48:29.042630 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 7 01:48:29.049830 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 7 01:48:29.054471 systemd[1]: Mounted media.mount - External Media Directory. Mar 7 01:48:29.058576 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 7 01:48:29.063912 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 7 01:48:29.069944 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 7 01:48:29.074602 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 7 01:48:29.080070 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 01:48:29.085737 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 7 01:48:29.085897 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 7 01:48:29.091165 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:48:29.091313 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:48:29.096577 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 7 01:48:29.096734 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 7 01:48:29.101577 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:48:29.101719 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:48:29.107096 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 7 01:48:29.107262 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 7 01:48:29.112588 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:48:29.112826 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:48:29.117837 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 7 01:48:29.123621 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 7 01:48:29.129059 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 7 01:48:29.134946 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 01:48:29.150843 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 7 01:48:29.160509 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 7 01:48:29.169549 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 7 01:48:29.174341 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 7 01:48:29.214559 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 7 01:48:29.220379 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 7 01:48:29.225461 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 01:48:29.226551 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 7 01:48:29.231176 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 01:48:29.232202 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 01:48:29.242037 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 7 01:48:29.261645 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 7 01:48:29.270334 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 7 01:48:29.275990 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 7 01:48:29.282100 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 7 01:48:29.291621 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 7 01:48:29.297509 systemd-journald[1303]: Time spent on flushing to /var/log/journal/f284eccc0e2c463db1f6ea526d021a7c is 43.122ms for 890 entries. Mar 7 01:48:29.297509 systemd-journald[1303]: System Journal (/var/log/journal/f284eccc0e2c463db1f6ea526d021a7c) is 11.8M, max 2.6G, 2.6G free. Mar 7 01:48:29.385110 systemd-journald[1303]: Received client request to flush runtime journal. Mar 7 01:48:29.385161 systemd-journald[1303]: /var/log/journal/f284eccc0e2c463db1f6ea526d021a7c/system.journal: Realtime clock jumped backwards relative to last journal entry, rotating. Mar 7 01:48:29.385183 systemd-journald[1303]: Rotating system journal. Mar 7 01:48:29.297884 udevadm[1346]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Mar 7 01:48:29.390954 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 7 01:48:29.398277 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 01:48:29.419068 systemd-tmpfiles[1344]: ACLs are not supported, ignoring. Mar 7 01:48:29.419086 systemd-tmpfiles[1344]: ACLs are not supported, ignoring. Mar 7 01:48:29.424898 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 7 01:48:29.438580 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 7 01:48:29.567270 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 7 01:48:29.578689 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 7 01:48:29.594958 systemd-tmpfiles[1365]: ACLs are not supported, ignoring. Mar 7 01:48:29.594975 systemd-tmpfiles[1365]: ACLs are not supported, ignoring. Mar 7 01:48:29.600802 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 01:48:29.909365 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 7 01:48:29.919597 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 01:48:29.950192 systemd-udevd[1371]: Using default interface naming scheme 'v255'. Mar 7 01:48:30.104510 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 01:48:30.118579 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 7 01:48:30.171614 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 7 01:48:30.180555 systemd[1]: Found device dev-ttyAMA0.device - /dev/ttyAMA0. Mar 7 01:48:30.237122 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 7 01:48:30.284522 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#196 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 7 01:48:30.284771 kernel: mousedev: PS/2 mouse device common for all mice Mar 7 01:48:30.315387 kernel: hv_vmbus: registering driver hv_balloon Mar 7 01:48:30.315500 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Mar 7 01:48:30.319381 kernel: hv_balloon: Memory hot add disabled on ARM64 Mar 7 01:48:30.327578 kernel: hv_vmbus: registering driver hyperv_fb Mar 7 01:48:30.339702 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Mar 7 01:48:30.339775 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Mar 7 01:48:30.347356 kernel: Console: switching to colour dummy device 80x25 Mar 7 01:48:30.350786 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:30.355463 kernel: Console: switching to colour frame buffer device 128x48 Mar 7 01:48:30.369036 systemd-networkd[1378]: lo: Link UP Mar 7 01:48:30.369045 systemd-networkd[1378]: lo: Gained carrier Mar 7 01:48:30.371851 systemd-networkd[1378]: Enumeration completed Mar 7 01:48:30.373424 systemd-networkd[1378]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:48:30.373430 systemd-networkd[1378]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:48:30.373736 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 7 01:48:30.385661 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 7 01:48:30.394304 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 01:48:30.394569 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:30.410583 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 01:48:30.436453 kernel: mlx5_core 4556:00:02.0 enP17750s1: Link up Mar 7 01:48:30.463176 kernel: hv_netvsc 7ced8d79-84eb-7ced-8d79-84eb7ced8d79 eth0: Data path switched to VF: enP17750s1 Mar 7 01:48:30.462721 systemd-networkd[1378]: enP17750s1: Link UP Mar 7 01:48:30.462811 systemd-networkd[1378]: eth0: Link UP Mar 7 01:48:30.462815 systemd-networkd[1378]: eth0: Gained carrier Mar 7 01:48:30.462829 systemd-networkd[1378]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:48:30.468069 systemd-networkd[1378]: enP17750s1: Gained carrier Mar 7 01:48:30.473496 systemd-networkd[1378]: eth0: DHCPv4 address 10.200.20.15/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 7 01:48:30.497503 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1381) Mar 7 01:48:30.554267 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Mar 7 01:48:30.614523 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 7 01:48:30.626553 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 7 01:48:30.703449 lvm[1462]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 7 01:48:30.728332 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 7 01:48:30.734011 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 7 01:48:30.744695 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 7 01:48:30.748740 lvm[1465]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 7 01:48:30.768847 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 7 01:48:30.774859 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 7 01:48:30.780036 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 7 01:48:30.780059 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 7 01:48:30.784342 systemd[1]: Reached target machines.target - Containers. Mar 7 01:48:30.790078 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 7 01:48:30.800553 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 7 01:48:30.806600 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 7 01:48:30.811025 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:48:30.812594 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 7 01:48:30.819191 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 7 01:48:30.828434 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 7 01:48:30.837120 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 7 01:48:30.869561 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 7 01:48:30.871215 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 7 01:48:30.898458 kernel: loop0: detected capacity change from 0 to 31320 Mar 7 01:48:30.905635 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 7 01:48:30.932794 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 01:48:31.521639 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 7 01:48:31.599459 kernel: loop1: detected capacity change from 0 to 209336 Mar 7 01:48:31.696580 kernel: loop2: detected capacity change from 0 to 114432 Mar 7 01:48:31.936591 systemd-networkd[1378]: eth0: Gained IPv6LL Mar 7 01:48:31.939356 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 7 01:48:32.107460 kernel: loop3: detected capacity change from 0 to 114328 Mar 7 01:48:32.513470 kernel: loop4: detected capacity change from 0 to 31320 Mar 7 01:48:32.529459 kernel: loop5: detected capacity change from 0 to 209336 Mar 7 01:48:32.546509 kernel: loop6: detected capacity change from 0 to 114432 Mar 7 01:48:32.558532 kernel: loop7: detected capacity change from 0 to 114328 Mar 7 01:48:32.566601 (sd-merge)[1492]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Mar 7 01:48:32.567013 (sd-merge)[1492]: Merged extensions into '/usr'. Mar 7 01:48:32.570423 systemd[1]: Reloading requested from client PID 1473 ('systemd-sysext') (unit systemd-sysext.service)... Mar 7 01:48:32.570435 systemd[1]: Reloading... Mar 7 01:48:32.630460 zram_generator::config[1522]: No configuration found. Mar 7 01:48:32.762850 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:48:32.832883 systemd[1]: Reloading finished in 262 ms. Mar 7 01:48:32.850247 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 7 01:48:32.861752 systemd[1]: Starting ensure-sysext.service... Mar 7 01:48:32.867676 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 7 01:48:32.875778 systemd[1]: Reloading requested from client PID 1580 ('systemctl') (unit ensure-sysext.service)... Mar 7 01:48:32.875792 systemd[1]: Reloading... Mar 7 01:48:32.897860 systemd-tmpfiles[1581]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 7 01:48:32.898676 systemd-tmpfiles[1581]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 7 01:48:32.899793 systemd-tmpfiles[1581]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 7 01:48:32.900288 systemd-tmpfiles[1581]: ACLs are not supported, ignoring. Mar 7 01:48:32.900511 systemd-tmpfiles[1581]: ACLs are not supported, ignoring. Mar 7 01:48:32.904873 systemd-tmpfiles[1581]: Detected autofs mount point /boot during canonicalization of boot. Mar 7 01:48:32.904977 systemd-tmpfiles[1581]: Skipping /boot Mar 7 01:48:32.913795 systemd-tmpfiles[1581]: Detected autofs mount point /boot during canonicalization of boot. Mar 7 01:48:32.913896 systemd-tmpfiles[1581]: Skipping /boot Mar 7 01:48:32.957570 zram_generator::config[1611]: No configuration found. Mar 7 01:48:33.066524 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:48:33.140277 systemd[1]: Reloading finished in 264 ms. Mar 7 01:48:33.155363 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 01:48:33.169711 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 7 01:48:33.183599 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 7 01:48:33.194585 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 7 01:48:33.208643 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 7 01:48:33.220585 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 7 01:48:33.233804 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:48:33.236212 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:48:33.252732 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:48:33.273774 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:48:33.280236 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:48:33.281426 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 7 01:48:33.289073 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:48:33.289225 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:48:33.296386 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:48:33.296702 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:48:33.302753 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:48:33.304604 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:48:33.319692 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 7 01:48:33.320455 augenrules[1702]: No rules Mar 7 01:48:33.327905 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:48:33.332759 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:48:33.339711 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:48:33.346655 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:48:33.351406 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:48:33.352267 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 7 01:48:33.363720 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:48:33.363897 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:48:33.369486 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:48:33.369638 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:48:33.375869 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:48:33.376075 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:48:33.385589 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 01:48:33.397670 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 01:48:33.407465 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 7 01:48:33.427648 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 01:48:33.434372 systemd-resolved[1679]: Positive Trust Anchors: Mar 7 01:48:33.434388 systemd-resolved[1679]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 7 01:48:33.434419 systemd-resolved[1679]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 7 01:48:33.434802 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 01:48:33.439765 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 01:48:33.439937 systemd[1]: Reached target time-set.target - System Time Set. Mar 7 01:48:33.445186 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 01:48:33.445339 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 01:48:33.451224 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 7 01:48:33.451369 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 7 01:48:33.457735 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 01:48:33.457891 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 01:48:33.465188 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 01:48:33.465377 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 01:48:33.473826 systemd[1]: Finished ensure-sysext.service. Mar 7 01:48:33.480665 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 01:48:33.480731 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 01:48:33.491709 systemd-resolved[1679]: Using system hostname 'ci-4081.3.6-n-e8b1fc37af'. Mar 7 01:48:33.493272 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 7 01:48:33.498833 systemd[1]: Reached target network.target - Network. Mar 7 01:48:33.502701 systemd[1]: Reached target network-online.target - Network is Online. Mar 7 01:48:33.507471 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 7 01:48:34.004283 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 7 01:48:34.010452 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 7 01:48:36.597177 ldconfig[1469]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 7 01:48:36.620861 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 7 01:48:36.630638 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 7 01:48:36.643795 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 7 01:48:36.648785 systemd[1]: Reached target sysinit.target - System Initialization. Mar 7 01:48:36.653527 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 7 01:48:36.658674 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 7 01:48:36.664418 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 7 01:48:36.669035 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 7 01:48:36.674352 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 7 01:48:36.680275 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 7 01:48:36.680304 systemd[1]: Reached target paths.target - Path Units. Mar 7 01:48:36.684251 systemd[1]: Reached target timers.target - Timer Units. Mar 7 01:48:36.690542 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 7 01:48:36.696851 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 7 01:48:36.701995 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 7 01:48:36.707367 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 7 01:48:36.712524 systemd[1]: Reached target sockets.target - Socket Units. Mar 7 01:48:36.716463 systemd[1]: Reached target basic.target - Basic System. Mar 7 01:48:36.720427 systemd[1]: System is tainted: cgroupsv1 Mar 7 01:48:36.720475 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 7 01:48:36.720495 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 7 01:48:36.722676 systemd[1]: Starting chronyd.service - NTP client/server... Mar 7 01:48:36.729557 systemd[1]: Starting containerd.service - containerd container runtime... Mar 7 01:48:36.736859 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 7 01:48:36.755951 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 7 01:48:36.764560 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 7 01:48:36.768840 (chronyd)[1750]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Mar 7 01:48:36.772617 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 7 01:48:36.774096 jq[1757]: false Mar 7 01:48:36.778786 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 7 01:48:36.778826 systemd[1]: hv_fcopy_daemon.service - Hyper-V FCOPY daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_fcopy). Mar 7 01:48:36.784965 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Mar 7 01:48:36.786881 KVP[1760]: KVP starting; pid is:1760 Mar 7 01:48:36.791776 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Mar 7 01:48:36.793564 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:48:36.801182 chronyd[1764]: chronyd version 4.5 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) Mar 7 01:48:36.811610 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 7 01:48:36.818000 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 7 01:48:36.829836 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 7 01:48:36.835588 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 7 01:48:36.842753 KVP[1760]: KVP LIC Version: 3.1 Mar 7 01:48:36.843671 kernel: hv_utils: KVP IC version 4.0 Mar 7 01:48:36.848076 extend-filesystems[1758]: Found loop4 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found loop5 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found loop6 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found loop7 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda1 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda2 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda3 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found usr Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda4 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda6 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda7 Mar 7 01:48:36.860812 extend-filesystems[1758]: Found sda9 Mar 7 01:48:36.860812 extend-filesystems[1758]: Checking size of /dev/sda9 Mar 7 01:48:36.855899 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 7 01:48:36.854290 chronyd[1764]: Timezone right/UTC failed leap second check, ignoring Mar 7 01:48:37.027459 extend-filesystems[1758]: Old size kept for /dev/sda9 Mar 7 01:48:37.027459 extend-filesystems[1758]: Found sr0 Mar 7 01:48:36.878621 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 7 01:48:36.854490 chronyd[1764]: Loaded seccomp filter (level 2) Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.033 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.033 INFO Fetch successful Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.034 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.048 INFO Fetch successful Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.048 INFO Fetching http://168.63.129.16/machine/9a0bac3d-7f1b-483a-ac74-2273dcd77e17/12b293fa%2Df16b%2D4c11%2D8424%2D392234b4d21a.%5Fci%2D4081.3.6%2Dn%2De8b1fc37af?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.048 INFO Fetch successful Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.048 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Mar 7 01:48:37.069569 coreos-metadata[1752]: Mar 07 01:48:37.057 INFO Fetch successful Mar 7 01:48:36.886905 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 7 01:48:36.932435 dbus-daemon[1755]: [system] SELinux support is enabled Mar 7 01:48:36.889584 systemd[1]: Starting update-engine.service - Update Engine... Mar 7 01:48:36.900537 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 7 01:48:36.921129 systemd[1]: Started chronyd.service - NTP client/server. Mar 7 01:48:37.070977 update_engine[1787]: I20260307 01:48:37.007236 1787 main.cc:92] Flatcar Update Engine starting Mar 7 01:48:37.070977 update_engine[1787]: I20260307 01:48:37.018909 1787 update_check_scheduler.cc:74] Next update check in 4m58s Mar 7 01:48:36.941090 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 7 01:48:37.071248 jq[1788]: true Mar 7 01:48:36.962736 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 7 01:48:36.962959 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 7 01:48:36.963194 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 7 01:48:36.963383 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 7 01:48:37.081850 jq[1816]: true Mar 7 01:48:36.972744 systemd[1]: motdgen.service: Deactivated successfully. Mar 7 01:48:36.972943 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 7 01:48:37.005781 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 7 01:48:37.024200 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 7 01:48:37.024475 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 7 01:48:37.063001 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 7 01:48:37.063029 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 7 01:48:37.070130 (ntainerd)[1817]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 7 01:48:37.072710 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 7 01:48:37.072732 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 7 01:48:37.086521 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1810) Mar 7 01:48:37.097615 systemd[1]: Started update-engine.service - Update Engine. Mar 7 01:48:37.121431 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 7 01:48:37.122514 systemd-logind[1784]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 7 01:48:37.123583 systemd-logind[1784]: New seat seat0. Mar 7 01:48:37.126697 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 7 01:48:37.137036 systemd[1]: Started systemd-logind.service - User Login Management. Mar 7 01:48:37.162393 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 7 01:48:37.170604 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 7 01:48:37.208553 tar[1815]: linux-arm64/LICENSE Mar 7 01:48:37.208553 tar[1815]: linux-arm64/helm Mar 7 01:48:37.277837 bash[1883]: Updated "/home/core/.ssh/authorized_keys" Mar 7 01:48:37.279307 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 7 01:48:37.291654 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 7 01:48:37.515903 locksmithd[1855]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 7 01:48:37.747774 tar[1815]: linux-arm64/README.md Mar 7 01:48:37.767849 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 7 01:48:37.911771 containerd[1817]: time="2026-03-07T01:48:37.911692520Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Mar 7 01:48:37.961369 containerd[1817]: time="2026-03-07T01:48:37.960634200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:48:37.970030 containerd[1817]: time="2026-03-07T01:48:37.969987840Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:48:37.970150 containerd[1817]: time="2026-03-07T01:48:37.970135600Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 7 01:48:37.970224 containerd[1817]: time="2026-03-07T01:48:37.970210640Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 7 01:48:37.970428 containerd[1817]: time="2026-03-07T01:48:37.970408800Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 7 01:48:37.970997 containerd[1817]: time="2026-03-07T01:48:37.970976680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 7 01:48:37.971350 containerd[1817]: time="2026-03-07T01:48:37.971329200Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:48:37.971938 containerd[1817]: time="2026-03-07T01:48:37.971829200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:48:37.973413 containerd[1817]: time="2026-03-07T01:48:37.973110360Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:48:37.973413 containerd[1817]: time="2026-03-07T01:48:37.973135880Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 7 01:48:37.973413 containerd[1817]: time="2026-03-07T01:48:37.973151320Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:48:37.973413 containerd[1817]: time="2026-03-07T01:48:37.973161720Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 7 01:48:37.973413 containerd[1817]: time="2026-03-07T01:48:37.973288360Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:48:37.974343 containerd[1817]: time="2026-03-07T01:48:37.974312640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 7 01:48:37.974622 containerd[1817]: time="2026-03-07T01:48:37.974593360Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 01:48:37.974764 containerd[1817]: time="2026-03-07T01:48:37.974704960Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 7 01:48:37.975459 containerd[1817]: time="2026-03-07T01:48:37.975372480Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 7 01:48:37.975686 containerd[1817]: time="2026-03-07T01:48:37.975541920Z" level=info msg="metadata content store policy set" policy=shared Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999255240Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999341680Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999361480Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999377920Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999392880Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999576840Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999876240Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999978400Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:37.999994400Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:38.000007000Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:38.000021440Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:38.000034000Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:38.000045800Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000473 containerd[1817]: time="2026-03-07T01:48:38.000060760Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000075520Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000089040Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000101880Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000113840Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000133640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000149240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000161120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000174080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000186040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000198680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000211000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000224720Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000243960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.000801 containerd[1817]: time="2026-03-07T01:48:38.000258120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000274000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000286400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000298400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000316200Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000336480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000348960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000359520Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000406080Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 7 01:48:38.001079 containerd[1817]: time="2026-03-07T01:48:38.000423760Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 7 01:48:38.002456 containerd[1817]: time="2026-03-07T01:48:38.000434080Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 7 01:48:38.002456 containerd[1817]: time="2026-03-07T01:48:38.001910240Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 7 01:48:38.002456 containerd[1817]: time="2026-03-07T01:48:38.001928640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.002456 containerd[1817]: time="2026-03-07T01:48:38.001948880Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 7 01:48:38.002456 containerd[1817]: time="2026-03-07T01:48:38.001966080Z" level=info msg="NRI interface is disabled by configuration." Mar 7 01:48:38.002456 containerd[1817]: time="2026-03-07T01:48:38.001978040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 7 01:48:38.003765 containerd[1817]: time="2026-03-07T01:48:38.002857240Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 7 01:48:38.003765 containerd[1817]: time="2026-03-07T01:48:38.002924560Z" level=info msg="Connect containerd service" Mar 7 01:48:38.003765 containerd[1817]: time="2026-03-07T01:48:38.002957280Z" level=info msg="using legacy CRI server" Mar 7 01:48:38.003765 containerd[1817]: time="2026-03-07T01:48:38.002963960Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 7 01:48:38.003765 containerd[1817]: time="2026-03-07T01:48:38.003044240Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 7 01:48:38.005279 containerd[1817]: time="2026-03-07T01:48:38.005131680Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 7 01:48:38.006072 containerd[1817]: time="2026-03-07T01:48:38.005914320Z" level=info msg="Start subscribing containerd event" Mar 7 01:48:38.006072 containerd[1817]: time="2026-03-07T01:48:38.005975040Z" level=info msg="Start recovering state" Mar 7 01:48:38.006818 containerd[1817]: time="2026-03-07T01:48:38.006240880Z" level=info msg="Start event monitor" Mar 7 01:48:38.006916 containerd[1817]: time="2026-03-07T01:48:38.006901320Z" level=info msg="Start snapshots syncer" Mar 7 01:48:38.007030 containerd[1817]: time="2026-03-07T01:48:38.006996880Z" level=info msg="Start cni network conf syncer for default" Mar 7 01:48:38.007150 containerd[1817]: time="2026-03-07T01:48:38.007136840Z" level=info msg="Start streaming server" Mar 7 01:48:38.007287 containerd[1817]: time="2026-03-07T01:48:38.006300280Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 7 01:48:38.009462 containerd[1817]: time="2026-03-07T01:48:38.007391360Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 7 01:48:38.009462 containerd[1817]: time="2026-03-07T01:48:38.008074320Z" level=info msg="containerd successfully booted in 0.097423s" Mar 7 01:48:38.007563 systemd[1]: Started containerd.service - containerd container runtime. Mar 7 01:48:38.091424 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:48:38.096821 (kubelet)[1914]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:48:38.525342 kubelet[1914]: E0307 01:48:38.525302 1914 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:48:38.531139 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:48:38.531413 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:48:38.941136 sshd_keygen[1791]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 7 01:48:38.960806 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 7 01:48:38.972970 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 7 01:48:38.978819 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Mar 7 01:48:38.983871 systemd[1]: issuegen.service: Deactivated successfully. Mar 7 01:48:38.984200 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 7 01:48:39.001048 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 7 01:48:39.019914 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 7 01:48:39.027655 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Mar 7 01:48:39.035729 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 7 01:48:39.041591 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Mar 7 01:48:39.047301 systemd[1]: Reached target getty.target - Login Prompts. Mar 7 01:48:39.052078 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 7 01:48:39.056999 systemd[1]: Startup finished in 14.948s (kernel) + 14.411s (userspace) = 29.359s. Mar 7 01:48:39.376453 login[1949]: pam_lastlog(login:session): file /var/log/lastlog is locked/write, retrying Mar 7 01:48:39.378424 login[1950]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:39.389043 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 7 01:48:39.389078 systemd-logind[1784]: New session 2 of user core. Mar 7 01:48:39.401625 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 7 01:48:39.428771 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 7 01:48:39.436857 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 7 01:48:39.454993 (systemd)[1959]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 7 01:48:39.624013 systemd[1959]: Queued start job for default target default.target. Mar 7 01:48:39.624671 systemd[1959]: Created slice app.slice - User Application Slice. Mar 7 01:48:39.624864 systemd[1959]: Reached target paths.target - Paths. Mar 7 01:48:39.624879 systemd[1959]: Reached target timers.target - Timers. Mar 7 01:48:39.632566 systemd[1959]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 7 01:48:39.638974 systemd[1959]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 7 01:48:39.639026 systemd[1959]: Reached target sockets.target - Sockets. Mar 7 01:48:39.639038 systemd[1959]: Reached target basic.target - Basic System. Mar 7 01:48:39.639081 systemd[1959]: Reached target default.target - Main User Target. Mar 7 01:48:39.639103 systemd[1959]: Startup finished in 178ms. Mar 7 01:48:39.639231 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 7 01:48:39.647761 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 7 01:48:40.376803 login[1949]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:40.383943 systemd-logind[1784]: New session 1 of user core. Mar 7 01:48:40.388642 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 7 01:48:40.862365 waagent[1947]: 2026-03-07T01:48:40.862274Z INFO Daemon Daemon Azure Linux Agent Version: 2.9.1.1 Mar 7 01:48:40.867154 waagent[1947]: 2026-03-07T01:48:40.867087Z INFO Daemon Daemon OS: flatcar 4081.3.6 Mar 7 01:48:40.871032 waagent[1947]: 2026-03-07T01:48:40.870984Z INFO Daemon Daemon Python: 3.11.9 Mar 7 01:48:40.874700 waagent[1947]: 2026-03-07T01:48:40.874629Z INFO Daemon Daemon Run daemon Mar 7 01:48:40.877990 waagent[1947]: 2026-03-07T01:48:40.877949Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4081.3.6' Mar 7 01:48:40.885407 waagent[1947]: 2026-03-07T01:48:40.885331Z INFO Daemon Daemon Using waagent for provisioning Mar 7 01:48:40.890026 waagent[1947]: 2026-03-07T01:48:40.889975Z INFO Daemon Daemon Activate resource disk Mar 7 01:48:40.893891 waagent[1947]: 2026-03-07T01:48:40.893847Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Mar 7 01:48:40.904112 waagent[1947]: 2026-03-07T01:48:40.904058Z INFO Daemon Daemon Found device: None Mar 7 01:48:40.907733 waagent[1947]: 2026-03-07T01:48:40.907690Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Mar 7 01:48:40.914340 waagent[1947]: 2026-03-07T01:48:40.914303Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Mar 7 01:48:40.925433 waagent[1947]: 2026-03-07T01:48:40.925379Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 7 01:48:40.930131 waagent[1947]: 2026-03-07T01:48:40.930087Z INFO Daemon Daemon Running default provisioning handler Mar 7 01:48:40.941669 waagent[1947]: 2026-03-07T01:48:40.941611Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Mar 7 01:48:40.953178 waagent[1947]: 2026-03-07T01:48:40.953127Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Mar 7 01:48:40.960512 waagent[1947]: 2026-03-07T01:48:40.960469Z INFO Daemon Daemon cloud-init is enabled: False Mar 7 01:48:40.964506 waagent[1947]: 2026-03-07T01:48:40.964472Z INFO Daemon Daemon Copying ovf-env.xml Mar 7 01:48:41.080606 waagent[1947]: 2026-03-07T01:48:41.080520Z INFO Daemon Daemon Successfully mounted dvd Mar 7 01:48:41.093941 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Mar 7 01:48:41.095885 waagent[1947]: 2026-03-07T01:48:41.095528Z INFO Daemon Daemon Detect protocol endpoint Mar 7 01:48:41.099546 waagent[1947]: 2026-03-07T01:48:41.099497Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 7 01:48:41.103960 waagent[1947]: 2026-03-07T01:48:41.103920Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Mar 7 01:48:41.109069 waagent[1947]: 2026-03-07T01:48:41.109033Z INFO Daemon Daemon Test for route to 168.63.129.16 Mar 7 01:48:41.113464 waagent[1947]: 2026-03-07T01:48:41.113361Z INFO Daemon Daemon Route to 168.63.129.16 exists Mar 7 01:48:41.117529 waagent[1947]: 2026-03-07T01:48:41.117486Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Mar 7 01:48:41.149736 waagent[1947]: 2026-03-07T01:48:41.149685Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Mar 7 01:48:41.155297 waagent[1947]: 2026-03-07T01:48:41.155267Z INFO Daemon Daemon Wire protocol version:2012-11-30 Mar 7 01:48:41.159717 waagent[1947]: 2026-03-07T01:48:41.159661Z INFO Daemon Daemon Server preferred version:2015-04-05 Mar 7 01:48:41.348484 waagent[1947]: 2026-03-07T01:48:41.348216Z INFO Daemon Daemon Initializing goal state during protocol detection Mar 7 01:48:41.353314 waagent[1947]: 2026-03-07T01:48:41.353266Z INFO Daemon Daemon Forcing an update of the goal state. Mar 7 01:48:41.360330 waagent[1947]: 2026-03-07T01:48:41.360284Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Mar 7 01:48:41.378662 waagent[1947]: 2026-03-07T01:48:41.378562Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.179 Mar 7 01:48:41.383131 waagent[1947]: 2026-03-07T01:48:41.383086Z INFO Daemon Mar 7 01:48:41.385389 waagent[1947]: 2026-03-07T01:48:41.385352Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: f6f71227-1687-4e15-a165-d8970a67b368 eTag: 10250755892245017877 source: Fabric] Mar 7 01:48:41.394341 waagent[1947]: 2026-03-07T01:48:41.394297Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Mar 7 01:48:41.399661 waagent[1947]: 2026-03-07T01:48:41.399617Z INFO Daemon Mar 7 01:48:41.401813 waagent[1947]: 2026-03-07T01:48:41.401778Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Mar 7 01:48:41.410188 waagent[1947]: 2026-03-07T01:48:41.410154Z INFO Daemon Daemon Downloading artifacts profile blob Mar 7 01:48:41.483132 waagent[1947]: 2026-03-07T01:48:41.483063Z INFO Daemon Downloaded certificate {'thumbprint': 'B86010E5594EA753FB72424EA386DB72632A1BB0', 'hasPrivateKey': True} Mar 7 01:48:41.490891 waagent[1947]: 2026-03-07T01:48:41.490845Z INFO Daemon Fetch goal state completed Mar 7 01:48:41.500884 waagent[1947]: 2026-03-07T01:48:41.500841Z INFO Daemon Daemon Starting provisioning Mar 7 01:48:41.505023 waagent[1947]: 2026-03-07T01:48:41.504974Z INFO Daemon Daemon Handle ovf-env.xml. Mar 7 01:48:41.508926 waagent[1947]: 2026-03-07T01:48:41.508889Z INFO Daemon Daemon Set hostname [ci-4081.3.6-n-e8b1fc37af] Mar 7 01:48:41.530683 waagent[1947]: 2026-03-07T01:48:41.530615Z INFO Daemon Daemon Publish hostname [ci-4081.3.6-n-e8b1fc37af] Mar 7 01:48:41.535639 waagent[1947]: 2026-03-07T01:48:41.535586Z INFO Daemon Daemon Examine /proc/net/route for primary interface Mar 7 01:48:41.540425 waagent[1947]: 2026-03-07T01:48:41.540381Z INFO Daemon Daemon Primary interface is [eth0] Mar 7 01:48:41.568138 systemd-networkd[1378]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 01:48:41.568149 systemd-networkd[1378]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 01:48:41.568194 systemd-networkd[1378]: eth0: DHCP lease lost Mar 7 01:48:41.569615 waagent[1947]: 2026-03-07T01:48:41.569550Z INFO Daemon Daemon Create user account if not exists Mar 7 01:48:41.573777 waagent[1947]: 2026-03-07T01:48:41.573735Z INFO Daemon Daemon User core already exists, skip useradd Mar 7 01:48:41.577846 systemd-networkd[1378]: eth0: DHCPv6 lease lost Mar 7 01:48:41.578397 waagent[1947]: 2026-03-07T01:48:41.578346Z INFO Daemon Daemon Configure sudoer Mar 7 01:48:41.581927 waagent[1947]: 2026-03-07T01:48:41.581882Z INFO Daemon Daemon Configure sshd Mar 7 01:48:41.585394 waagent[1947]: 2026-03-07T01:48:41.585349Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Mar 7 01:48:41.596097 waagent[1947]: 2026-03-07T01:48:41.596044Z INFO Daemon Daemon Deploy ssh public key. Mar 7 01:48:41.604494 systemd-networkd[1378]: eth0: DHCPv4 address 10.200.20.15/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 7 01:48:42.693596 waagent[1947]: 2026-03-07T01:48:42.693522Z INFO Daemon Daemon Provisioning complete Mar 7 01:48:42.708317 waagent[1947]: 2026-03-07T01:48:42.708267Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Mar 7 01:48:42.713204 waagent[1947]: 2026-03-07T01:48:42.713157Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Mar 7 01:48:42.720655 waagent[1947]: 2026-03-07T01:48:42.720596Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.9.1.1 is the most current agent Mar 7 01:48:42.850247 waagent[2012]: 2026-03-07T01:48:42.849621Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.9.1.1) Mar 7 01:48:42.850247 waagent[2012]: 2026-03-07T01:48:42.849770Z INFO ExtHandler ExtHandler OS: flatcar 4081.3.6 Mar 7 01:48:42.850247 waagent[2012]: 2026-03-07T01:48:42.849822Z INFO ExtHandler ExtHandler Python: 3.11.9 Mar 7 01:48:42.902651 waagent[2012]: 2026-03-07T01:48:42.902573Z INFO ExtHandler ExtHandler Distro: flatcar-4081.3.6; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.9; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Mar 7 01:48:42.902968 waagent[2012]: 2026-03-07T01:48:42.902929Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 7 01:48:42.903113 waagent[2012]: 2026-03-07T01:48:42.903077Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 7 01:48:42.910824 waagent[2012]: 2026-03-07T01:48:42.910765Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Mar 7 01:48:42.916122 waagent[2012]: 2026-03-07T01:48:42.916076Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.179 Mar 7 01:48:42.916737 waagent[2012]: 2026-03-07T01:48:42.916691Z INFO ExtHandler Mar 7 01:48:42.916919 waagent[2012]: 2026-03-07T01:48:42.916881Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: d80c8f58-1495-4e50-9475-8fa865044d9d eTag: 10250755892245017877 source: Fabric] Mar 7 01:48:42.917312 waagent[2012]: 2026-03-07T01:48:42.917268Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Mar 7 01:48:42.918687 waagent[2012]: 2026-03-07T01:48:42.917977Z INFO ExtHandler Mar 7 01:48:42.918687 waagent[2012]: 2026-03-07T01:48:42.918064Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Mar 7 01:48:42.921480 waagent[2012]: 2026-03-07T01:48:42.921129Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Mar 7 01:48:42.995612 waagent[2012]: 2026-03-07T01:48:42.995467Z INFO ExtHandler Downloaded certificate {'thumbprint': 'B86010E5594EA753FB72424EA386DB72632A1BB0', 'hasPrivateKey': True} Mar 7 01:48:42.996082 waagent[2012]: 2026-03-07T01:48:42.996032Z INFO ExtHandler Fetch goal state completed Mar 7 01:48:43.010102 waagent[2012]: 2026-03-07T01:48:43.010044Z INFO ExtHandler ExtHandler WALinuxAgent-2.9.1.1 running as process 2012 Mar 7 01:48:43.010252 waagent[2012]: 2026-03-07T01:48:43.010217Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Mar 7 01:48:43.011893 waagent[2012]: 2026-03-07T01:48:43.011843Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '4081.3.6', '', 'Flatcar Container Linux by Kinvolk'] Mar 7 01:48:43.012262 waagent[2012]: 2026-03-07T01:48:43.012223Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Mar 7 01:48:43.078728 waagent[2012]: 2026-03-07T01:48:43.078680Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Mar 7 01:48:43.078936 waagent[2012]: 2026-03-07T01:48:43.078896Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Mar 7 01:48:43.084790 waagent[2012]: 2026-03-07T01:48:43.084754Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Mar 7 01:48:43.091217 systemd[1]: Reloading requested from client PID 2025 ('systemctl') (unit waagent.service)... Mar 7 01:48:43.091477 systemd[1]: Reloading... Mar 7 01:48:43.161519 zram_generator::config[2060]: No configuration found. Mar 7 01:48:43.286809 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:48:43.364307 systemd[1]: Reloading finished in 272 ms. Mar 7 01:48:43.383641 waagent[2012]: 2026-03-07T01:48:43.382796Z INFO ExtHandler ExtHandler Executing systemctl daemon-reload for setting up waagent-network-setup.service Mar 7 01:48:43.390541 systemd[1]: Reloading requested from client PID 2119 ('systemctl') (unit waagent.service)... Mar 7 01:48:43.390554 systemd[1]: Reloading... Mar 7 01:48:43.465483 zram_generator::config[2156]: No configuration found. Mar 7 01:48:43.570065 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:48:43.643886 systemd[1]: Reloading finished in 253 ms. Mar 7 01:48:43.667723 waagent[2012]: 2026-03-07T01:48:43.667643Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Mar 7 01:48:43.667845 waagent[2012]: 2026-03-07T01:48:43.667810Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Mar 7 01:48:44.124659 waagent[2012]: 2026-03-07T01:48:44.124570Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Mar 7 01:48:44.125215 waagent[2012]: 2026-03-07T01:48:44.125163Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [True], cgroups enabled [False], python supported: [True] Mar 7 01:48:44.126031 waagent[2012]: 2026-03-07T01:48:44.125972Z INFO ExtHandler ExtHandler Starting env monitor service. Mar 7 01:48:44.126524 waagent[2012]: 2026-03-07T01:48:44.126382Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Mar 7 01:48:44.126957 waagent[2012]: 2026-03-07T01:48:44.126848Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Mar 7 01:48:44.127050 waagent[2012]: 2026-03-07T01:48:44.126952Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Mar 7 01:48:44.128154 waagent[2012]: 2026-03-07T01:48:44.127392Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 7 01:48:44.128154 waagent[2012]: 2026-03-07T01:48:44.127514Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 7 01:48:44.128154 waagent[2012]: 2026-03-07T01:48:44.127719Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Mar 7 01:48:44.128154 waagent[2012]: 2026-03-07T01:48:44.127881Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Mar 7 01:48:44.128154 waagent[2012]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Mar 7 01:48:44.128154 waagent[2012]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Mar 7 01:48:44.128154 waagent[2012]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Mar 7 01:48:44.128154 waagent[2012]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Mar 7 01:48:44.128154 waagent[2012]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 7 01:48:44.128154 waagent[2012]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 7 01:48:44.128706 waagent[2012]: 2026-03-07T01:48:44.128502Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Mar 7 01:48:44.128774 waagent[2012]: 2026-03-07T01:48:44.128721Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Mar 7 01:48:44.129188 waagent[2012]: 2026-03-07T01:48:44.129143Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Mar 7 01:48:44.129530 waagent[2012]: 2026-03-07T01:48:44.129487Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 7 01:48:44.130785 waagent[2012]: 2026-03-07T01:48:44.130732Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 7 01:48:44.130946 waagent[2012]: 2026-03-07T01:48:44.130900Z INFO EnvHandler ExtHandler Configure routes Mar 7 01:48:44.131010 waagent[2012]: 2026-03-07T01:48:44.130980Z INFO EnvHandler ExtHandler Gateway:None Mar 7 01:48:44.131057 waagent[2012]: 2026-03-07T01:48:44.131032Z INFO EnvHandler ExtHandler Routes:None Mar 7 01:48:44.134602 waagent[2012]: 2026-03-07T01:48:44.134556Z INFO ExtHandler ExtHandler Mar 7 01:48:44.134690 waagent[2012]: 2026-03-07T01:48:44.134657Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: ab1cec96-b516-4252-884c-8e9caaf796d1 correlation 5848b290-6746-4cdc-afc5-326b201900c7 created: 2026-03-07T01:47:36.012863Z] Mar 7 01:48:44.135599 waagent[2012]: 2026-03-07T01:48:44.135544Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Mar 7 01:48:44.137206 waagent[2012]: 2026-03-07T01:48:44.137157Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 2 ms] Mar 7 01:48:44.176574 waagent[2012]: 2026-03-07T01:48:44.176515Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.9.1.1 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 30F90340-4FEB-483B-B66B-D0848286AF13;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 0] Mar 7 01:48:44.196949 waagent[2012]: 2026-03-07T01:48:44.196499Z INFO MonitorHandler ExtHandler Network interfaces: Mar 7 01:48:44.196949 waagent[2012]: Executing ['ip', '-a', '-o', 'link']: Mar 7 01:48:44.196949 waagent[2012]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Mar 7 01:48:44.196949 waagent[2012]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:79:84:eb brd ff:ff:ff:ff:ff:ff Mar 7 01:48:44.196949 waagent[2012]: 3: enP17750s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:79:84:eb brd ff:ff:ff:ff:ff:ff\ altname enP17750p0s2 Mar 7 01:48:44.196949 waagent[2012]: Executing ['ip', '-4', '-a', '-o', 'address']: Mar 7 01:48:44.196949 waagent[2012]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Mar 7 01:48:44.196949 waagent[2012]: 2: eth0 inet 10.200.20.15/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Mar 7 01:48:44.196949 waagent[2012]: Executing ['ip', '-6', '-a', '-o', 'address']: Mar 7 01:48:44.196949 waagent[2012]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Mar 7 01:48:44.196949 waagent[2012]: 2: eth0 inet6 fe80::7eed:8dff:fe79:84eb/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Mar 7 01:48:44.241702 waagent[2012]: 2026-03-07T01:48:44.241643Z INFO EnvHandler ExtHandler Successfully added Azure fabric firewall rules. Current Firewall rules: Mar 7 01:48:44.241702 waagent[2012]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 7 01:48:44.241702 waagent[2012]: pkts bytes target prot opt in out source destination Mar 7 01:48:44.241702 waagent[2012]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 7 01:48:44.241702 waagent[2012]: pkts bytes target prot opt in out source destination Mar 7 01:48:44.241702 waagent[2012]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 7 01:48:44.241702 waagent[2012]: pkts bytes target prot opt in out source destination Mar 7 01:48:44.241702 waagent[2012]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 7 01:48:44.241702 waagent[2012]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 7 01:48:44.241702 waagent[2012]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 7 01:48:44.244965 waagent[2012]: 2026-03-07T01:48:44.244913Z INFO EnvHandler ExtHandler Current Firewall rules: Mar 7 01:48:44.244965 waagent[2012]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 7 01:48:44.244965 waagent[2012]: pkts bytes target prot opt in out source destination Mar 7 01:48:44.244965 waagent[2012]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 7 01:48:44.244965 waagent[2012]: pkts bytes target prot opt in out source destination Mar 7 01:48:44.244965 waagent[2012]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 7 01:48:44.244965 waagent[2012]: pkts bytes target prot opt in out source destination Mar 7 01:48:44.244965 waagent[2012]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 7 01:48:44.244965 waagent[2012]: 4 594 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 7 01:48:44.244965 waagent[2012]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 7 01:48:44.245574 waagent[2012]: 2026-03-07T01:48:44.245421Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Mar 7 01:48:48.610319 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 7 01:48:48.618714 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:48:48.797621 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:48:48.801812 (kubelet)[2255]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:48:48.837929 kubelet[2255]: E0307 01:48:48.837862 2255 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:48:48.841408 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:48:48.841613 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:48:52.016346 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 7 01:48:52.026661 systemd[1]: Started sshd@0-10.200.20.15:22-10.200.16.10:33770.service - OpenSSH per-connection server daemon (10.200.16.10:33770). Mar 7 01:48:52.580870 sshd[2263]: Accepted publickey for core from 10.200.16.10 port 33770 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:48:52.581735 sshd[2263]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:52.585805 systemd-logind[1784]: New session 3 of user core. Mar 7 01:48:52.595673 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 7 01:48:53.012666 systemd[1]: Started sshd@1-10.200.20.15:22-10.200.16.10:33786.service - OpenSSH per-connection server daemon (10.200.16.10:33786). Mar 7 01:48:53.498293 sshd[2268]: Accepted publickey for core from 10.200.16.10 port 33786 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:48:53.499610 sshd[2268]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:53.503786 systemd-logind[1784]: New session 4 of user core. Mar 7 01:48:53.513699 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 7 01:48:53.851721 sshd[2268]: pam_unix(sshd:session): session closed for user core Mar 7 01:48:53.855794 systemd[1]: sshd@1-10.200.20.15:22-10.200.16.10:33786.service: Deactivated successfully. Mar 7 01:48:53.857800 systemd-logind[1784]: Session 4 logged out. Waiting for processes to exit. Mar 7 01:48:53.858261 systemd[1]: session-4.scope: Deactivated successfully. Mar 7 01:48:53.859503 systemd-logind[1784]: Removed session 4. Mar 7 01:48:53.923719 systemd[1]: Started sshd@2-10.200.20.15:22-10.200.16.10:33792.service - OpenSSH per-connection server daemon (10.200.16.10:33792). Mar 7 01:48:54.367735 sshd[2276]: Accepted publickey for core from 10.200.16.10 port 33792 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:48:54.369053 sshd[2276]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:54.372480 systemd-logind[1784]: New session 5 of user core. Mar 7 01:48:54.379741 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 7 01:48:54.688097 sshd[2276]: pam_unix(sshd:session): session closed for user core Mar 7 01:48:54.691024 systemd-logind[1784]: Session 5 logged out. Waiting for processes to exit. Mar 7 01:48:54.693838 systemd[1]: sshd@2-10.200.20.15:22-10.200.16.10:33792.service: Deactivated successfully. Mar 7 01:48:54.696031 systemd[1]: session-5.scope: Deactivated successfully. Mar 7 01:48:54.697321 systemd-logind[1784]: Removed session 5. Mar 7 01:48:54.782659 systemd[1]: Started sshd@3-10.200.20.15:22-10.200.16.10:33808.service - OpenSSH per-connection server daemon (10.200.16.10:33808). Mar 7 01:48:55.273090 sshd[2284]: Accepted publickey for core from 10.200.16.10 port 33808 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:48:55.273905 sshd[2284]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:55.277973 systemd-logind[1784]: New session 6 of user core. Mar 7 01:48:55.285687 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 7 01:48:55.628840 sshd[2284]: pam_unix(sshd:session): session closed for user core Mar 7 01:48:55.632536 systemd-logind[1784]: Session 6 logged out. Waiting for processes to exit. Mar 7 01:48:55.633186 systemd[1]: sshd@3-10.200.20.15:22-10.200.16.10:33808.service: Deactivated successfully. Mar 7 01:48:55.635068 systemd[1]: session-6.scope: Deactivated successfully. Mar 7 01:48:55.636327 systemd-logind[1784]: Removed session 6. Mar 7 01:48:55.724687 systemd[1]: Started sshd@4-10.200.20.15:22-10.200.16.10:33812.service - OpenSSH per-connection server daemon (10.200.16.10:33812). Mar 7 01:48:56.211463 sshd[2292]: Accepted publickey for core from 10.200.16.10 port 33812 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:48:56.212711 sshd[2292]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:56.217521 systemd-logind[1784]: New session 7 of user core. Mar 7 01:48:56.222769 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 7 01:48:56.631338 sudo[2296]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 7 01:48:56.631646 sudo[2296]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:48:56.677266 sudo[2296]: pam_unix(sudo:session): session closed for user root Mar 7 01:48:56.755751 sshd[2292]: pam_unix(sshd:session): session closed for user core Mar 7 01:48:56.761962 systemd[1]: sshd@4-10.200.20.15:22-10.200.16.10:33812.service: Deactivated successfully. Mar 7 01:48:56.762593 systemd-logind[1784]: Session 7 logged out. Waiting for processes to exit. Mar 7 01:48:56.765345 systemd[1]: session-7.scope: Deactivated successfully. Mar 7 01:48:56.766435 systemd-logind[1784]: Removed session 7. Mar 7 01:48:56.846968 systemd[1]: Started sshd@5-10.200.20.15:22-10.200.16.10:33816.service - OpenSSH per-connection server daemon (10.200.16.10:33816). Mar 7 01:48:57.335061 sshd[2301]: Accepted publickey for core from 10.200.16.10 port 33816 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:48:57.335907 sshd[2301]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:57.339451 systemd-logind[1784]: New session 8 of user core. Mar 7 01:48:57.346663 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 7 01:48:57.610835 sudo[2306]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 7 01:48:57.611550 sudo[2306]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:48:57.614759 sudo[2306]: pam_unix(sudo:session): session closed for user root Mar 7 01:48:57.619272 sudo[2305]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Mar 7 01:48:57.619622 sudo[2305]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:48:57.633710 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Mar 7 01:48:57.635110 auditctl[2309]: No rules Mar 7 01:48:57.636036 systemd[1]: audit-rules.service: Deactivated successfully. Mar 7 01:48:57.638665 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Mar 7 01:48:57.640891 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 7 01:48:57.669433 augenrules[2328]: No rules Mar 7 01:48:57.670673 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 7 01:48:57.671959 sudo[2305]: pam_unix(sudo:session): session closed for user root Mar 7 01:48:57.749461 sshd[2301]: pam_unix(sshd:session): session closed for user core Mar 7 01:48:57.754801 systemd[1]: sshd@5-10.200.20.15:22-10.200.16.10:33816.service: Deactivated successfully. Mar 7 01:48:57.757511 systemd[1]: session-8.scope: Deactivated successfully. Mar 7 01:48:57.757539 systemd-logind[1784]: Session 8 logged out. Waiting for processes to exit. Mar 7 01:48:57.758681 systemd-logind[1784]: Removed session 8. Mar 7 01:48:57.833639 systemd[1]: Started sshd@6-10.200.20.15:22-10.200.16.10:33828.service - OpenSSH per-connection server daemon (10.200.16.10:33828). Mar 7 01:48:58.317462 sshd[2337]: Accepted publickey for core from 10.200.16.10 port 33828 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:48:58.318267 sshd[2337]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:48:58.322667 systemd-logind[1784]: New session 9 of user core. Mar 7 01:48:58.331674 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 7 01:48:58.591343 sudo[2341]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 7 01:48:58.591644 sudo[2341]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 01:48:58.860253 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 7 01:48:58.865599 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:48:59.290646 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:48:59.295220 (kubelet)[2364]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:48:59.329769 kubelet[2364]: E0307 01:48:59.329716 2364 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:48:59.332096 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:48:59.332248 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:49:00.087691 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 7 01:49:00.087869 (dockerd)[2377]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 7 01:49:00.642493 chronyd[1764]: Selected source PHC0 Mar 7 01:49:00.749304 dockerd[2377]: time="2026-03-07T01:49:00.749207856Z" level=info msg="Starting up" Mar 7 01:49:01.516274 dockerd[2377]: time="2026-03-07T01:49:01.516201855Z" level=info msg="Loading containers: start." Mar 7 01:49:01.753468 kernel: Initializing XFRM netlink socket Mar 7 01:49:01.924358 systemd-networkd[1378]: docker0: Link UP Mar 7 01:49:01.955628 dockerd[2377]: time="2026-03-07T01:49:01.954833433Z" level=info msg="Loading containers: done." Mar 7 01:49:01.966798 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1420334411-merged.mount: Deactivated successfully. Mar 7 01:49:01.978481 dockerd[2377]: time="2026-03-07T01:49:01.978349099Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 7 01:49:01.978588 dockerd[2377]: time="2026-03-07T01:49:01.978506899Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Mar 7 01:49:01.978681 dockerd[2377]: time="2026-03-07T01:49:01.978658139Z" level=info msg="Daemon has completed initialization" Mar 7 01:49:02.055820 dockerd[2377]: time="2026-03-07T01:49:02.055117704Z" level=info msg="API listen on /run/docker.sock" Mar 7 01:49:02.056095 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 7 01:49:02.436446 containerd[1817]: time="2026-03-07T01:49:02.436397963Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\"" Mar 7 01:49:03.280619 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount9610826.mount: Deactivated successfully. Mar 7 01:49:04.766005 containerd[1817]: time="2026-03-07T01:49:04.765948900Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:04.768851 containerd[1817]: time="2026-03-07T01:49:04.768628943Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.9: active requests=0, bytes read=27390174" Mar 7 01:49:04.771849 containerd[1817]: time="2026-03-07T01:49:04.771821466Z" level=info msg="ImageCreate event name:\"sha256:6dbc3c6e88c8bca1294fa5fafe73dbe01fb58d40e562dbfc8b8b4195940270c8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:04.777320 containerd[1817]: time="2026-03-07T01:49:04.777291552Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:04.778933 containerd[1817]: time="2026-03-07T01:49:04.778343674Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.9\" with image id \"sha256:6dbc3c6e88c8bca1294fa5fafe73dbe01fb58d40e562dbfc8b8b4195940270c8\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\", size \"27386773\" in 2.341905951s" Mar 7 01:49:04.778933 containerd[1817]: time="2026-03-07T01:49:04.778380714Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\" returns image reference \"sha256:6dbc3c6e88c8bca1294fa5fafe73dbe01fb58d40e562dbfc8b8b4195940270c8\"" Mar 7 01:49:04.779339 containerd[1817]: time="2026-03-07T01:49:04.779245795Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\"" Mar 7 01:49:06.609417 containerd[1817]: time="2026-03-07T01:49:06.609368901Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:06.612252 containerd[1817]: time="2026-03-07T01:49:06.612065584Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.9: active requests=0, bytes read=23552106" Mar 7 01:49:06.615117 containerd[1817]: time="2026-03-07T01:49:06.615092467Z" level=info msg="ImageCreate event name:\"sha256:c58be92c40cc41b6c83c361b92110b587104386f93c5b7a9fc66dffdd1523d17\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:06.620963 containerd[1817]: time="2026-03-07T01:49:06.620929234Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:06.623006 containerd[1817]: time="2026-03-07T01:49:06.622971756Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.9\" with image id \"sha256:c58be92c40cc41b6c83c361b92110b587104386f93c5b7a9fc66dffdd1523d17\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\", size \"25136510\" in 1.843541121s" Mar 7 01:49:06.623067 containerd[1817]: time="2026-03-07T01:49:06.623007076Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\" returns image reference \"sha256:c58be92c40cc41b6c83c361b92110b587104386f93c5b7a9fc66dffdd1523d17\"" Mar 7 01:49:06.623565 containerd[1817]: time="2026-03-07T01:49:06.623541076Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\"" Mar 7 01:49:08.056618 containerd[1817]: time="2026-03-07T01:49:08.055580582Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:08.061839 containerd[1817]: time="2026-03-07T01:49:08.061800549Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.9: active requests=0, bytes read=18301305" Mar 7 01:49:08.067153 containerd[1817]: time="2026-03-07T01:49:08.065750633Z" level=info msg="ImageCreate event name:\"sha256:5dcd4a0c93d95bd92241ba240a130ffbde67814e2b417a13c25738a7b0204e95\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:08.073541 containerd[1817]: time="2026-03-07T01:49:08.073512802Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:08.074782 containerd[1817]: time="2026-03-07T01:49:08.074270443Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.9\" with image id \"sha256:5dcd4a0c93d95bd92241ba240a130ffbde67814e2b417a13c25738a7b0204e95\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\", size \"19885727\" in 1.450698167s" Mar 7 01:49:08.075315 containerd[1817]: time="2026-03-07T01:49:08.075295164Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\" returns image reference \"sha256:5dcd4a0c93d95bd92241ba240a130ffbde67814e2b417a13c25738a7b0204e95\"" Mar 7 01:49:08.075869 containerd[1817]: time="2026-03-07T01:49:08.075815204Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\"" Mar 7 01:49:09.360237 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 7 01:49:09.363359 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount380781696.mount: Deactivated successfully. Mar 7 01:49:09.371657 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:49:09.522670 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:49:09.524624 (kubelet)[2595]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:49:09.603918 kubelet[2595]: E0307 01:49:09.603855 2595 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:49:09.606028 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:49:09.606164 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:49:10.221266 containerd[1817]: time="2026-03-07T01:49:10.221209336Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:10.229658 containerd[1817]: time="2026-03-07T01:49:10.229622610Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.9: active requests=0, bytes read=28148870" Mar 7 01:49:10.237860 containerd[1817]: time="2026-03-07T01:49:10.237830443Z" level=info msg="ImageCreate event name:\"sha256:fb4f3cb8cccaec5975890c2ee802236a557e3f108da9c3c66ebec335ac73dcc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:10.252953 containerd[1817]: time="2026-03-07T01:49:10.252898224Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:10.253872 containerd[1817]: time="2026-03-07T01:49:10.253538386Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.9\" with image id \"sha256:fb4f3cb8cccaec5975890c2ee802236a557e3f108da9c3c66ebec335ac73dcc9\", repo tag \"registry.k8s.io/kube-proxy:v1.33.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\", size \"28147889\" in 2.177539182s" Mar 7 01:49:10.253872 containerd[1817]: time="2026-03-07T01:49:10.253571306Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\" returns image reference \"sha256:fb4f3cb8cccaec5975890c2ee802236a557e3f108da9c3c66ebec335ac73dcc9\"" Mar 7 01:49:10.254426 containerd[1817]: time="2026-03-07T01:49:10.254233589Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Mar 7 01:49:10.921032 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1219125141.mount: Deactivated successfully. Mar 7 01:49:12.270789 containerd[1817]: time="2026-03-07T01:49:12.270740925Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:12.273381 containerd[1817]: time="2026-03-07T01:49:12.273349449Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=19152117" Mar 7 01:49:12.277219 containerd[1817]: time="2026-03-07T01:49:12.277192935Z" level=info msg="ImageCreate event name:\"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:12.283114 containerd[1817]: time="2026-03-07T01:49:12.283071663Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:12.284306 containerd[1817]: time="2026-03-07T01:49:12.284158505Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"19148915\" in 2.029892676s" Mar 7 01:49:12.284306 containerd[1817]: time="2026-03-07T01:49:12.284192425Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\"" Mar 7 01:49:12.284823 containerd[1817]: time="2026-03-07T01:49:12.284797585Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 7 01:49:12.848609 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount880966183.mount: Deactivated successfully. Mar 7 01:49:12.870156 containerd[1817]: time="2026-03-07T01:49:12.870110665Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:12.874015 containerd[1817]: time="2026-03-07T01:49:12.873750590Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Mar 7 01:49:12.877424 containerd[1817]: time="2026-03-07T01:49:12.877170475Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:12.883594 containerd[1817]: time="2026-03-07T01:49:12.883550804Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:12.884629 containerd[1817]: time="2026-03-07T01:49:12.884238485Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 599.411699ms" Mar 7 01:49:12.884629 containerd[1817]: time="2026-03-07T01:49:12.884271045Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Mar 7 01:49:12.884912 containerd[1817]: time="2026-03-07T01:49:12.884889726Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Mar 7 01:49:13.559529 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4243562519.mount: Deactivated successfully. Mar 7 01:49:16.282593 containerd[1817]: time="2026-03-07T01:49:16.282538079Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:16.287191 containerd[1817]: time="2026-03-07T01:49:16.287160486Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=21885780" Mar 7 01:49:16.295769 containerd[1817]: time="2026-03-07T01:49:16.295254537Z" level=info msg="ImageCreate event name:\"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:16.302712 containerd[1817]: time="2026-03-07T01:49:16.302676988Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:16.303888 containerd[1817]: time="2026-03-07T01:49:16.303856750Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"21882972\" in 3.418523943s" Mar 7 01:49:16.303992 containerd[1817]: time="2026-03-07T01:49:16.303974750Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\"" Mar 7 01:49:18.454455 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Mar 7 01:49:19.611081 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 7 01:49:19.616721 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:49:19.930598 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:49:19.934105 (kubelet)[2759]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 01:49:19.972415 kubelet[2759]: E0307 01:49:19.972348 2759 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 01:49:19.975649 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 01:49:19.975786 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 01:49:20.145190 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:49:20.154938 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:49:20.184046 systemd[1]: Reloading requested from client PID 2776 ('systemctl') (unit session-9.scope)... Mar 7 01:49:20.184059 systemd[1]: Reloading... Mar 7 01:49:20.286485 zram_generator::config[2816]: No configuration found. Mar 7 01:49:20.401401 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:49:20.477406 systemd[1]: Reloading finished in 293 ms. Mar 7 01:49:20.518666 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 7 01:49:20.518743 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 7 01:49:20.518984 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:49:20.529072 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:49:20.680631 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:49:20.690811 (kubelet)[2895]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 7 01:49:20.722885 kubelet[2895]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:49:20.722885 kubelet[2895]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 7 01:49:20.722885 kubelet[2895]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:49:20.723262 kubelet[2895]: I0307 01:49:20.722929 2895 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 7 01:49:21.797471 kubelet[2895]: I0307 01:49:21.796783 2895 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 7 01:49:21.797471 kubelet[2895]: I0307 01:49:21.796811 2895 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 7 01:49:21.797471 kubelet[2895]: I0307 01:49:21.797030 2895 server.go:956] "Client rotation is on, will bootstrap in background" Mar 7 01:49:21.814205 kubelet[2895]: E0307 01:49:21.814150 2895 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.200.20.15:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 7 01:49:21.815633 kubelet[2895]: I0307 01:49:21.815607 2895 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 7 01:49:21.821200 kubelet[2895]: E0307 01:49:21.821045 2895 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 7 01:49:21.821200 kubelet[2895]: I0307 01:49:21.821072 2895 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 7 01:49:21.827458 kubelet[2895]: I0307 01:49:21.827396 2895 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 7 01:49:21.827796 kubelet[2895]: I0307 01:49:21.827760 2895 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 7 01:49:21.827950 kubelet[2895]: I0307 01:49:21.827791 2895 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.6-n-e8b1fc37af","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 7 01:49:21.828038 kubelet[2895]: I0307 01:49:21.827954 2895 topology_manager.go:138] "Creating topology manager with none policy" Mar 7 01:49:21.828038 kubelet[2895]: I0307 01:49:21.827963 2895 container_manager_linux.go:303] "Creating device plugin manager" Mar 7 01:49:21.828094 kubelet[2895]: I0307 01:49:21.828081 2895 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:49:21.830866 kubelet[2895]: I0307 01:49:21.830848 2895 kubelet.go:480] "Attempting to sync node with API server" Mar 7 01:49:21.830911 kubelet[2895]: I0307 01:49:21.830871 2895 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 7 01:49:21.830911 kubelet[2895]: I0307 01:49:21.830895 2895 kubelet.go:386] "Adding apiserver pod source" Mar 7 01:49:21.830911 kubelet[2895]: I0307 01:49:21.830908 2895 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 7 01:49:21.836740 kubelet[2895]: E0307 01:49:21.836419 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.200.20.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 7 01:49:21.836740 kubelet[2895]: E0307 01:49:21.836708 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.200.20.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-e8b1fc37af&limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 7 01:49:21.837324 kubelet[2895]: I0307 01:49:21.837302 2895 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 7 01:49:21.837855 kubelet[2895]: I0307 01:49:21.837839 2895 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 7 01:49:21.837910 kubelet[2895]: W0307 01:49:21.837898 2895 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 7 01:49:21.841354 kubelet[2895]: I0307 01:49:21.841335 2895 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 7 01:49:21.841416 kubelet[2895]: I0307 01:49:21.841378 2895 server.go:1289] "Started kubelet" Mar 7 01:49:21.842485 kubelet[2895]: I0307 01:49:21.841494 2895 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 7 01:49:21.842485 kubelet[2895]: I0307 01:49:21.842236 2895 server.go:317] "Adding debug handlers to kubelet server" Mar 7 01:49:21.843493 kubelet[2895]: I0307 01:49:21.843155 2895 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 7 01:49:21.843571 kubelet[2895]: I0307 01:49:21.843432 2895 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 7 01:49:21.844688 kubelet[2895]: E0307 01:49:21.843731 2895 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.15:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.15:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.6-n-e8b1fc37af.189a6c029f42116d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.6-n-e8b1fc37af,UID:ci-4081.3.6-n-e8b1fc37af,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.6-n-e8b1fc37af,},FirstTimestamp:2026-03-07 01:49:21.841353069 +0000 UTC m=+1.146838434,LastTimestamp:2026-03-07 01:49:21.841353069 +0000 UTC m=+1.146838434,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.6-n-e8b1fc37af,}" Mar 7 01:49:21.845931 kubelet[2895]: I0307 01:49:21.845909 2895 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 7 01:49:21.847216 kubelet[2895]: E0307 01:49:21.847197 2895 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 7 01:49:21.848351 kubelet[2895]: I0307 01:49:21.847489 2895 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 7 01:49:21.850134 kubelet[2895]: E0307 01:49:21.850104 2895 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" Mar 7 01:49:21.850464 kubelet[2895]: I0307 01:49:21.850240 2895 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 7 01:49:21.850464 kubelet[2895]: I0307 01:49:21.850424 2895 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 7 01:49:21.850590 kubelet[2895]: I0307 01:49:21.850581 2895 reconciler.go:26] "Reconciler: start to sync state" Mar 7 01:49:21.851590 kubelet[2895]: E0307 01:49:21.851569 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.200.20.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 7 01:49:21.852486 kubelet[2895]: I0307 01:49:21.852472 2895 factory.go:223] Registration of the containerd container factory successfully Mar 7 01:49:21.852569 kubelet[2895]: I0307 01:49:21.852561 2895 factory.go:223] Registration of the systemd container factory successfully Mar 7 01:49:21.852706 kubelet[2895]: I0307 01:49:21.852691 2895 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 7 01:49:21.862549 kubelet[2895]: E0307 01:49:21.862509 2895 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-e8b1fc37af?timeout=10s\": dial tcp 10.200.20.15:6443: connect: connection refused" interval="200ms" Mar 7 01:49:21.865567 kubelet[2895]: I0307 01:49:21.865540 2895 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 7 01:49:21.866706 kubelet[2895]: I0307 01:49:21.866501 2895 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 7 01:49:21.866706 kubelet[2895]: I0307 01:49:21.866520 2895 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 7 01:49:21.866706 kubelet[2895]: I0307 01:49:21.866537 2895 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 7 01:49:21.866706 kubelet[2895]: I0307 01:49:21.866544 2895 kubelet.go:2436] "Starting kubelet main sync loop" Mar 7 01:49:21.866706 kubelet[2895]: E0307 01:49:21.866588 2895 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 7 01:49:21.874411 kubelet[2895]: E0307 01:49:21.874387 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.200.20.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 7 01:49:21.896759 kubelet[2895]: I0307 01:49:21.896740 2895 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 7 01:49:21.896759 kubelet[2895]: I0307 01:49:21.896754 2895 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 7 01:49:21.896857 kubelet[2895]: I0307 01:49:21.896771 2895 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:49:21.902414 kubelet[2895]: I0307 01:49:21.902389 2895 policy_none.go:49] "None policy: Start" Mar 7 01:49:21.902471 kubelet[2895]: I0307 01:49:21.902422 2895 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 7 01:49:21.902471 kubelet[2895]: I0307 01:49:21.902454 2895 state_mem.go:35] "Initializing new in-memory state store" Mar 7 01:49:21.909275 kubelet[2895]: E0307 01:49:21.909251 2895 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 7 01:49:21.909436 kubelet[2895]: I0307 01:49:21.909417 2895 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 7 01:49:21.909480 kubelet[2895]: I0307 01:49:21.909434 2895 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 7 01:49:21.910769 kubelet[2895]: I0307 01:49:21.910750 2895 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 7 01:49:21.911618 kubelet[2895]: E0307 01:49:21.911518 2895 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 7 01:49:21.911618 kubelet[2895]: E0307 01:49:21.911555 2895 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.6-n-e8b1fc37af\" not found" Mar 7 01:49:21.976662 kubelet[2895]: E0307 01:49:21.976391 2895 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:21.988642 kubelet[2895]: E0307 01:49:21.988623 2895 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:21.990416 kubelet[2895]: E0307 01:49:21.990394 2895 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.011511 kubelet[2895]: I0307 01:49:22.011486 2895 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.011842 kubelet[2895]: E0307 01:49:22.011815 2895 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.20.15:6443/api/v1/nodes\": dial tcp 10.200.20.15:6443: connect: connection refused" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.064259 kubelet[2895]: E0307 01:49:22.063703 2895 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-e8b1fc37af?timeout=10s\": dial tcp 10.200.20.15:6443: connect: connection refused" interval="400ms" Mar 7 01:49:22.152008 kubelet[2895]: I0307 01:49:22.151976 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6a585b3ec52cbbfb367010843964baa4-ca-certs\") pod \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" (UID: \"6a585b3ec52cbbfb367010843964baa4\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152008 kubelet[2895]: I0307 01:49:22.152015 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6a585b3ec52cbbfb367010843964baa4-k8s-certs\") pod \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" (UID: \"6a585b3ec52cbbfb367010843964baa4\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152157 kubelet[2895]: I0307 01:49:22.152033 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-ca-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152157 kubelet[2895]: I0307 01:49:22.152051 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152157 kubelet[2895]: I0307 01:49:22.152066 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/80095c29a85f6f15b80b1363d9e82d8f-kubeconfig\") pod \"kube-scheduler-ci-4081.3.6-n-e8b1fc37af\" (UID: \"80095c29a85f6f15b80b1363d9e82d8f\") " pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152157 kubelet[2895]: I0307 01:49:22.152081 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6a585b3ec52cbbfb367010843964baa4-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" (UID: \"6a585b3ec52cbbfb367010843964baa4\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152157 kubelet[2895]: I0307 01:49:22.152095 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152265 kubelet[2895]: I0307 01:49:22.152109 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.152265 kubelet[2895]: I0307 01:49:22.152125 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.213593 kubelet[2895]: I0307 01:49:22.213562 2895 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.213928 kubelet[2895]: E0307 01:49:22.213902 2895 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.20.15:6443/api/v1/nodes\": dial tcp 10.200.20.15:6443: connect: connection refused" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.277840 containerd[1817]: time="2026-03-07T01:49:22.277562999Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.6-n-e8b1fc37af,Uid:6a585b3ec52cbbfb367010843964baa4,Namespace:kube-system,Attempt:0,}" Mar 7 01:49:22.290170 containerd[1817]: time="2026-03-07T01:49:22.289946863Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.6-n-e8b1fc37af,Uid:0afdcccb48cf8cd82622e951bf7a70be,Namespace:kube-system,Attempt:0,}" Mar 7 01:49:22.291933 containerd[1817]: time="2026-03-07T01:49:22.291901107Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.6-n-e8b1fc37af,Uid:80095c29a85f6f15b80b1363d9e82d8f,Namespace:kube-system,Attempt:0,}" Mar 7 01:49:22.464184 kubelet[2895]: E0307 01:49:22.464136 2895 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-e8b1fc37af?timeout=10s\": dial tcp 10.200.20.15:6443: connect: connection refused" interval="800ms" Mar 7 01:49:22.616461 kubelet[2895]: I0307 01:49:22.616312 2895 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.616701 kubelet[2895]: E0307 01:49:22.616666 2895 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.20.15:6443/api/v1/nodes\": dial tcp 10.200.20.15:6443: connect: connection refused" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:22.704562 update_engine[1787]: I20260307 01:49:22.704489 1787 update_attempter.cc:509] Updating boot flags... Mar 7 01:49:22.752484 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (2940) Mar 7 01:49:22.853377 kubelet[2895]: E0307 01:49:22.853341 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.200.20.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 7 01:49:22.932928 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1919226291.mount: Deactivated successfully. Mar 7 01:49:22.962483 containerd[1817]: time="2026-03-07T01:49:22.962015612Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:49:22.965566 containerd[1817]: time="2026-03-07T01:49:22.965525579Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Mar 7 01:49:22.969072 containerd[1817]: time="2026-03-07T01:49:22.969035105Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:49:22.972793 containerd[1817]: time="2026-03-07T01:49:22.972030671Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:49:22.976991 containerd[1817]: time="2026-03-07T01:49:22.976958801Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 7 01:49:22.981467 containerd[1817]: time="2026-03-07T01:49:22.980425248Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:49:22.984288 containerd[1817]: time="2026-03-07T01:49:22.984255375Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 7 01:49:22.991807 containerd[1817]: time="2026-03-07T01:49:22.991759750Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 01:49:22.992781 containerd[1817]: time="2026-03-07T01:49:22.992743112Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 702.724609ms" Mar 7 01:49:22.994361 containerd[1817]: time="2026-03-07T01:49:22.994318115Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 716.673596ms" Mar 7 01:49:22.997390 containerd[1817]: time="2026-03-07T01:49:22.997343601Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 705.379814ms" Mar 7 01:49:23.071386 kubelet[2895]: E0307 01:49:23.071332 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.200.20.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 7 01:49:23.129031 kubelet[2895]: E0307 01:49:23.128976 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.200.20.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-e8b1fc37af&limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 7 01:49:23.265595 kubelet[2895]: E0307 01:49:23.265554 2895 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-e8b1fc37af?timeout=10s\": dial tcp 10.200.20.15:6443: connect: connection refused" interval="1.6s" Mar 7 01:49:23.420748 kubelet[2895]: I0307 01:49:23.420080 2895 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:23.420748 kubelet[2895]: E0307 01:49:23.420409 2895 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.20.15:6443/api/v1/nodes\": dial tcp 10.200.20.15:6443: connect: connection refused" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:23.434579 kubelet[2895]: E0307 01:49:23.434555 2895 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.200.20.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.20.15:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 7 01:49:23.616833 containerd[1817]: time="2026-03-07T01:49:23.616266566Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:49:23.616833 containerd[1817]: time="2026-03-07T01:49:23.616317086Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:49:23.616833 containerd[1817]: time="2026-03-07T01:49:23.616336006Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:23.616833 containerd[1817]: time="2026-03-07T01:49:23.616428926Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:23.617511 containerd[1817]: time="2026-03-07T01:49:23.616955607Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:49:23.617511 containerd[1817]: time="2026-03-07T01:49:23.617024847Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:49:23.617511 containerd[1817]: time="2026-03-07T01:49:23.617051688Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:23.619672 containerd[1817]: time="2026-03-07T01:49:23.619545492Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:23.622663 containerd[1817]: time="2026-03-07T01:49:23.622599418Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:49:23.622762 containerd[1817]: time="2026-03-07T01:49:23.622679779Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:49:23.622762 containerd[1817]: time="2026-03-07T01:49:23.622698019Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:23.622870 containerd[1817]: time="2026-03-07T01:49:23.622803179Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:23.690159 containerd[1817]: time="2026-03-07T01:49:23.689541265Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.6-n-e8b1fc37af,Uid:80095c29a85f6f15b80b1363d9e82d8f,Namespace:kube-system,Attempt:0,} returns sandbox id \"50ac5d24ac588264b3afff25b854a53ef883c6e967e38dbe67c23df4db35b7c8\"" Mar 7 01:49:23.696867 containerd[1817]: time="2026-03-07T01:49:23.696831276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.6-n-e8b1fc37af,Uid:0afdcccb48cf8cd82622e951bf7a70be,Namespace:kube-system,Attempt:0,} returns sandbox id \"0f142920976549481f33db158c59cc86c7f428c87c8415dbdf277fb8cf3f00fd\"" Mar 7 01:49:23.698944 containerd[1817]: time="2026-03-07T01:49:23.698914839Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.6-n-e8b1fc37af,Uid:6a585b3ec52cbbfb367010843964baa4,Namespace:kube-system,Attempt:0,} returns sandbox id \"048056d9aec39934bd0f24d9dedeca4e3ea6fbc5f498d358d9d7524f40cbbeca\"" Mar 7 01:49:23.714829 containerd[1817]: time="2026-03-07T01:49:23.714791863Z" level=info msg="CreateContainer within sandbox \"50ac5d24ac588264b3afff25b854a53ef883c6e967e38dbe67c23df4db35b7c8\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 7 01:49:23.724959 containerd[1817]: time="2026-03-07T01:49:23.724921638Z" level=info msg="CreateContainer within sandbox \"0f142920976549481f33db158c59cc86c7f428c87c8415dbdf277fb8cf3f00fd\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 7 01:49:23.730955 containerd[1817]: time="2026-03-07T01:49:23.730924088Z" level=info msg="CreateContainer within sandbox \"048056d9aec39934bd0f24d9dedeca4e3ea6fbc5f498d358d9d7524f40cbbeca\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 7 01:49:23.780745 containerd[1817]: time="2026-03-07T01:49:23.780569123Z" level=info msg="CreateContainer within sandbox \"50ac5d24ac588264b3afff25b854a53ef883c6e967e38dbe67c23df4db35b7c8\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"48bc3f4640696a60b5fb27a30581e6a379f7ba215797fae0affff471aa75deb8\"" Mar 7 01:49:23.781448 containerd[1817]: time="2026-03-07T01:49:23.781413884Z" level=info msg="StartContainer for \"48bc3f4640696a60b5fb27a30581e6a379f7ba215797fae0affff471aa75deb8\"" Mar 7 01:49:23.794417 containerd[1817]: time="2026-03-07T01:49:23.794378384Z" level=info msg="CreateContainer within sandbox \"048056d9aec39934bd0f24d9dedeca4e3ea6fbc5f498d358d9d7524f40cbbeca\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"53611855d57087a626f675850db7c2ae23d6d625e7f489a5c0da5f699dc95387\"" Mar 7 01:49:23.794837 containerd[1817]: time="2026-03-07T01:49:23.794814705Z" level=info msg="StartContainer for \"53611855d57087a626f675850db7c2ae23d6d625e7f489a5c0da5f699dc95387\"" Mar 7 01:49:23.803562 containerd[1817]: time="2026-03-07T01:49:23.802786437Z" level=info msg="CreateContainer within sandbox \"0f142920976549481f33db158c59cc86c7f428c87c8415dbdf277fb8cf3f00fd\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"95aef40e7b47ebe0631fbf63f14cfd458e0956e2f08066750f97d8288ef8fcb3\"" Mar 7 01:49:23.803649 containerd[1817]: time="2026-03-07T01:49:23.803601198Z" level=info msg="StartContainer for \"95aef40e7b47ebe0631fbf63f14cfd458e0956e2f08066750f97d8288ef8fcb3\"" Mar 7 01:49:23.852483 containerd[1817]: time="2026-03-07T01:49:23.852419033Z" level=info msg="StartContainer for \"48bc3f4640696a60b5fb27a30581e6a379f7ba215797fae0affff471aa75deb8\" returns successfully" Mar 7 01:49:23.892388 containerd[1817]: time="2026-03-07T01:49:23.892192733Z" level=info msg="StartContainer for \"53611855d57087a626f675850db7c2ae23d6d625e7f489a5c0da5f699dc95387\" returns successfully" Mar 7 01:49:23.905879 kubelet[2895]: E0307 01:49:23.905831 2895 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:23.949062 containerd[1817]: time="2026-03-07T01:49:23.948499819Z" level=info msg="StartContainer for \"95aef40e7b47ebe0631fbf63f14cfd458e0956e2f08066750f97d8288ef8fcb3\" returns successfully" Mar 7 01:49:24.913104 kubelet[2895]: E0307 01:49:24.912720 2895 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:24.920460 kubelet[2895]: E0307 01:49:24.919761 2895 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:24.920460 kubelet[2895]: E0307 01:49:24.919815 2895 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.023922 kubelet[2895]: I0307 01:49:25.023891 2895 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.698714 kubelet[2895]: E0307 01:49:25.698667 2895 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.6-n-e8b1fc37af\" not found" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.706679 kubelet[2895]: E0307 01:49:25.706567 2895 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4081.3.6-n-e8b1fc37af.189a6c029f42116d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.6-n-e8b1fc37af,UID:ci-4081.3.6-n-e8b1fc37af,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.6-n-e8b1fc37af,},FirstTimestamp:2026-03-07 01:49:21.841353069 +0000 UTC m=+1.146838434,LastTimestamp:2026-03-07 01:49:21.841353069 +0000 UTC m=+1.146838434,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.6-n-e8b1fc37af,}" Mar 7 01:49:25.741737 kubelet[2895]: I0307 01:49:25.741691 2895 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.741737 kubelet[2895]: E0307 01:49:25.741727 2895 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4081.3.6-n-e8b1fc37af\": node \"ci-4081.3.6-n-e8b1fc37af\" not found" Mar 7 01:49:25.765147 kubelet[2895]: E0307 01:49:25.765117 2895 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.6-n-e8b1fc37af\" not found" Mar 7 01:49:25.835620 kubelet[2895]: I0307 01:49:25.835588 2895 apiserver.go:52] "Watching apiserver" Mar 7 01:49:25.851597 kubelet[2895]: I0307 01:49:25.851571 2895 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 7 01:49:25.853931 kubelet[2895]: I0307 01:49:25.853778 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.871950 kubelet[2895]: E0307 01:49:25.871652 2895 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-e8b1fc37af\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.871950 kubelet[2895]: I0307 01:49:25.871676 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.876778 kubelet[2895]: E0307 01:49:25.876748 2895 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.876778 kubelet[2895]: I0307 01:49:25.876776 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.883416 kubelet[2895]: E0307 01:49:25.883388 2895 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.918036 kubelet[2895]: I0307 01:49:25.917823 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.918036 kubelet[2895]: I0307 01:49:25.917881 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.922178 kubelet[2895]: E0307 01:49:25.922153 2895 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:25.923474 kubelet[2895]: E0307 01:49:25.922496 2895 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:26.787187 kubelet[2895]: I0307 01:49:26.786488 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:26.795897 kubelet[2895]: I0307 01:49:26.795498 2895 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 7 01:49:26.920587 kubelet[2895]: I0307 01:49:26.918913 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:26.927943 kubelet[2895]: I0307 01:49:26.927702 2895 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 7 01:49:27.895845 systemd[1]: Reloading requested from client PID 3218 ('systemctl') (unit session-9.scope)... Mar 7 01:49:27.896127 systemd[1]: Reloading... Mar 7 01:49:27.983468 zram_generator::config[3256]: No configuration found. Mar 7 01:49:28.103835 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 01:49:28.138368 kubelet[2895]: I0307 01:49:28.138107 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.148372 kubelet[2895]: I0307 01:49:28.147564 2895 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 7 01:49:28.190602 systemd[1]: Reloading finished in 294 ms. Mar 7 01:49:28.217070 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:49:28.231276 systemd[1]: kubelet.service: Deactivated successfully. Mar 7 01:49:28.231627 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:49:28.242201 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 01:49:28.445609 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 01:49:28.448237 (kubelet)[3332]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 7 01:49:28.482796 kubelet[3332]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:49:28.483394 kubelet[3332]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 7 01:49:28.483394 kubelet[3332]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 01:49:28.483394 kubelet[3332]: I0307 01:49:28.483188 3332 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 7 01:49:28.492852 kubelet[3332]: I0307 01:49:28.492819 3332 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 7 01:49:28.492852 kubelet[3332]: I0307 01:49:28.492845 3332 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 7 01:49:28.493058 kubelet[3332]: I0307 01:49:28.493041 3332 server.go:956] "Client rotation is on, will bootstrap in background" Mar 7 01:49:28.494661 kubelet[3332]: I0307 01:49:28.494640 3332 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 7 01:49:28.499598 kubelet[3332]: I0307 01:49:28.499477 3332 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 7 01:49:28.503486 kubelet[3332]: E0307 01:49:28.503393 3332 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 7 01:49:28.503700 kubelet[3332]: I0307 01:49:28.503618 3332 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 7 01:49:28.507016 kubelet[3332]: I0307 01:49:28.506981 3332 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 7 01:49:28.507691 kubelet[3332]: I0307 01:49:28.507652 3332 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 7 01:49:28.507955 kubelet[3332]: I0307 01:49:28.507770 3332 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.6-n-e8b1fc37af","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 7 01:49:28.508194 kubelet[3332]: I0307 01:49:28.508121 3332 topology_manager.go:138] "Creating topology manager with none policy" Mar 7 01:49:28.508194 kubelet[3332]: I0307 01:49:28.508136 3332 container_manager_linux.go:303] "Creating device plugin manager" Mar 7 01:49:28.508545 kubelet[3332]: I0307 01:49:28.508284 3332 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:49:28.508545 kubelet[3332]: I0307 01:49:28.508455 3332 kubelet.go:480] "Attempting to sync node with API server" Mar 7 01:49:28.508545 kubelet[3332]: I0307 01:49:28.508471 3332 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 7 01:49:28.508545 kubelet[3332]: I0307 01:49:28.508493 3332 kubelet.go:386] "Adding apiserver pod source" Mar 7 01:49:28.508545 kubelet[3332]: I0307 01:49:28.508506 3332 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 7 01:49:28.526577 kubelet[3332]: I0307 01:49:28.525822 3332 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 7 01:49:28.526577 kubelet[3332]: I0307 01:49:28.526402 3332 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 7 01:49:28.531059 kubelet[3332]: I0307 01:49:28.530323 3332 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 7 01:49:28.531059 kubelet[3332]: I0307 01:49:28.530365 3332 server.go:1289] "Started kubelet" Mar 7 01:49:28.534188 kubelet[3332]: I0307 01:49:28.532116 3332 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 7 01:49:28.534188 kubelet[3332]: I0307 01:49:28.532172 3332 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 7 01:49:28.534188 kubelet[3332]: I0307 01:49:28.532415 3332 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 7 01:49:28.534188 kubelet[3332]: I0307 01:49:28.532479 3332 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 7 01:49:28.534188 kubelet[3332]: I0307 01:49:28.533182 3332 server.go:317] "Adding debug handlers to kubelet server" Mar 7 01:49:28.538598 kubelet[3332]: I0307 01:49:28.538576 3332 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 7 01:49:28.540532 kubelet[3332]: I0307 01:49:28.539571 3332 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 7 01:49:28.541504 kubelet[3332]: I0307 01:49:28.540944 3332 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 7 01:49:28.541504 kubelet[3332]: I0307 01:49:28.541055 3332 reconciler.go:26] "Reconciler: start to sync state" Mar 7 01:49:28.542534 kubelet[3332]: I0307 01:49:28.542515 3332 factory.go:223] Registration of the systemd container factory successfully Mar 7 01:49:28.542728 kubelet[3332]: I0307 01:49:28.542706 3332 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 7 01:49:28.543724 kubelet[3332]: I0307 01:49:28.543686 3332 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 7 01:49:28.547079 kubelet[3332]: I0307 01:49:28.547036 3332 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 7 01:49:28.547079 kubelet[3332]: I0307 01:49:28.547072 3332 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 7 01:49:28.547216 kubelet[3332]: I0307 01:49:28.547089 3332 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 7 01:49:28.547216 kubelet[3332]: I0307 01:49:28.547095 3332 kubelet.go:2436] "Starting kubelet main sync loop" Mar 7 01:49:28.547216 kubelet[3332]: E0307 01:49:28.547139 3332 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 7 01:49:28.549132 kubelet[3332]: I0307 01:49:28.549115 3332 factory.go:223] Registration of the containerd container factory successfully Mar 7 01:49:28.554712 kubelet[3332]: E0307 01:49:28.554668 3332 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 7 01:49:28.606055 kubelet[3332]: I0307 01:49:28.606028 3332 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 7 01:49:28.606055 kubelet[3332]: I0307 01:49:28.606047 3332 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 7 01:49:28.606201 kubelet[3332]: I0307 01:49:28.606071 3332 state_mem.go:36] "Initialized new in-memory state store" Mar 7 01:49:28.606223 kubelet[3332]: I0307 01:49:28.606202 3332 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 7 01:49:28.606259 kubelet[3332]: I0307 01:49:28.606211 3332 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 7 01:49:28.606259 kubelet[3332]: I0307 01:49:28.606240 3332 policy_none.go:49] "None policy: Start" Mar 7 01:49:28.606259 kubelet[3332]: I0307 01:49:28.606249 3332 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 7 01:49:28.606326 kubelet[3332]: I0307 01:49:28.606260 3332 state_mem.go:35] "Initializing new in-memory state store" Mar 7 01:49:28.606348 kubelet[3332]: I0307 01:49:28.606341 3332 state_mem.go:75] "Updated machine memory state" Mar 7 01:49:28.607388 kubelet[3332]: E0307 01:49:28.607366 3332 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 7 01:49:28.608730 kubelet[3332]: I0307 01:49:28.608603 3332 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 7 01:49:28.608730 kubelet[3332]: I0307 01:49:28.608621 3332 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 7 01:49:28.609034 kubelet[3332]: I0307 01:49:28.609021 3332 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 7 01:49:28.611756 kubelet[3332]: E0307 01:49:28.611072 3332 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 7 01:49:28.649261 kubelet[3332]: I0307 01:49:28.648694 3332 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.649261 kubelet[3332]: I0307 01:49:28.648723 3332 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.649261 kubelet[3332]: I0307 01:49:28.648875 3332 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.660110 kubelet[3332]: I0307 01:49:28.660075 3332 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 7 01:49:28.660227 kubelet[3332]: E0307 01:49:28.660138 3332 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-e8b1fc37af\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.660894 kubelet[3332]: I0307 01:49:28.660874 3332 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 7 01:49:28.660962 kubelet[3332]: E0307 01:49:28.660910 3332 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" already exists" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.661000 kubelet[3332]: I0307 01:49:28.660969 3332 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 7 01:49:28.661024 kubelet[3332]: E0307 01:49:28.660998 3332 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.716522 kubelet[3332]: I0307 01:49:28.715545 3332 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.735072 kubelet[3332]: I0307 01:49:28.735041 3332 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.735178 kubelet[3332]: I0307 01:49:28.735119 3332 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742698 kubelet[3332]: I0307 01:49:28.742667 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742805 kubelet[3332]: I0307 01:49:28.742705 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/80095c29a85f6f15b80b1363d9e82d8f-kubeconfig\") pod \"kube-scheduler-ci-4081.3.6-n-e8b1fc37af\" (UID: \"80095c29a85f6f15b80b1363d9e82d8f\") " pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742805 kubelet[3332]: I0307 01:49:28.742724 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6a585b3ec52cbbfb367010843964baa4-ca-certs\") pod \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" (UID: \"6a585b3ec52cbbfb367010843964baa4\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742805 kubelet[3332]: I0307 01:49:28.742738 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-ca-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742805 kubelet[3332]: I0307 01:49:28.742754 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742805 kubelet[3332]: I0307 01:49:28.742769 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742920 kubelet[3332]: I0307 01:49:28.742783 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0afdcccb48cf8cd82622e951bf7a70be-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.6-n-e8b1fc37af\" (UID: \"0afdcccb48cf8cd82622e951bf7a70be\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742920 kubelet[3332]: I0307 01:49:28.742798 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6a585b3ec52cbbfb367010843964baa4-k8s-certs\") pod \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" (UID: \"6a585b3ec52cbbfb367010843964baa4\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:28.742920 kubelet[3332]: I0307 01:49:28.742814 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6a585b3ec52cbbfb367010843964baa4-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" (UID: \"6a585b3ec52cbbfb367010843964baa4\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:29.510332 kubelet[3332]: I0307 01:49:29.510283 3332 apiserver.go:52] "Watching apiserver" Mar 7 01:49:29.541232 kubelet[3332]: I0307 01:49:29.541193 3332 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 7 01:49:29.581526 kubelet[3332]: I0307 01:49:29.581491 3332 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:29.595002 kubelet[3332]: I0307 01:49:29.594973 3332 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 7 01:49:29.595127 kubelet[3332]: E0307 01:49:29.595025 3332 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-e8b1fc37af\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" Mar 7 01:49:29.616774 kubelet[3332]: I0307 01:49:29.616692 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.6-n-e8b1fc37af" podStartSLOduration=3.616676924 podStartE2EDuration="3.616676924s" podCreationTimestamp="2026-03-07 01:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:49:29.603127773 +0000 UTC m=+1.150729809" watchObservedRunningTime="2026-03-07 01:49:29.616676924 +0000 UTC m=+1.164278960" Mar 7 01:49:29.629088 kubelet[3332]: I0307 01:49:29.629023 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.6-n-e8b1fc37af" podStartSLOduration=3.629007231 podStartE2EDuration="3.629007231s" podCreationTimestamp="2026-03-07 01:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:49:29.616914324 +0000 UTC m=+1.164516360" watchObservedRunningTime="2026-03-07 01:49:29.629007231 +0000 UTC m=+1.176609227" Mar 7 01:49:29.629245 kubelet[3332]: I0307 01:49:29.629137 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-e8b1fc37af" podStartSLOduration=1.629132152 podStartE2EDuration="1.629132152s" podCreationTimestamp="2026-03-07 01:49:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:49:29.628831831 +0000 UTC m=+1.176433867" watchObservedRunningTime="2026-03-07 01:49:29.629132152 +0000 UTC m=+1.176734188" Mar 7 01:49:34.499127 kubelet[3332]: I0307 01:49:34.499092 3332 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 7 01:49:34.499565 containerd[1817]: time="2026-03-07T01:49:34.499514825Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 7 01:49:34.500142 kubelet[3332]: I0307 01:49:34.499698 3332 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 7 01:49:35.479875 kubelet[3332]: I0307 01:49:35.479835 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb-kube-proxy\") pod \"kube-proxy-8lqsm\" (UID: \"6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb\") " pod="kube-system/kube-proxy-8lqsm" Mar 7 01:49:35.479875 kubelet[3332]: I0307 01:49:35.479876 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb-lib-modules\") pod \"kube-proxy-8lqsm\" (UID: \"6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb\") " pod="kube-system/kube-proxy-8lqsm" Mar 7 01:49:35.479875 kubelet[3332]: I0307 01:49:35.479895 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb-xtables-lock\") pod \"kube-proxy-8lqsm\" (UID: \"6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb\") " pod="kube-system/kube-proxy-8lqsm" Mar 7 01:49:35.479875 kubelet[3332]: I0307 01:49:35.479910 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b66dd\" (UniqueName: \"kubernetes.io/projected/6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb-kube-api-access-b66dd\") pod \"kube-proxy-8lqsm\" (UID: \"6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb\") " pod="kube-system/kube-proxy-8lqsm" Mar 7 01:49:35.590460 kubelet[3332]: E0307 01:49:35.590412 3332 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Mar 7 01:49:35.590823 kubelet[3332]: E0307 01:49:35.590583 3332 projected.go:194] Error preparing data for projected volume kube-api-access-b66dd for pod kube-system/kube-proxy-8lqsm: configmap "kube-root-ca.crt" not found Mar 7 01:49:35.590823 kubelet[3332]: E0307 01:49:35.590661 3332 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb-kube-api-access-b66dd podName:6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb nodeName:}" failed. No retries permitted until 2026-03-07 01:49:36.090640026 +0000 UTC m=+7.638242022 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-b66dd" (UniqueName: "kubernetes.io/projected/6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb-kube-api-access-b66dd") pod "kube-proxy-8lqsm" (UID: "6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb") : configmap "kube-root-ca.crt" not found Mar 7 01:49:35.882707 kubelet[3332]: I0307 01:49:35.882643 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sqrl\" (UniqueName: \"kubernetes.io/projected/856fa1da-0776-4eac-86d1-7571f0b78b29-kube-api-access-9sqrl\") pod \"tigera-operator-6bf85f8dd-c4kv7\" (UID: \"856fa1da-0776-4eac-86d1-7571f0b78b29\") " pod="tigera-operator/tigera-operator-6bf85f8dd-c4kv7" Mar 7 01:49:35.882707 kubelet[3332]: I0307 01:49:35.882713 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/856fa1da-0776-4eac-86d1-7571f0b78b29-var-lib-calico\") pod \"tigera-operator-6bf85f8dd-c4kv7\" (UID: \"856fa1da-0776-4eac-86d1-7571f0b78b29\") " pod="tigera-operator/tigera-operator-6bf85f8dd-c4kv7" Mar 7 01:49:36.088282 containerd[1817]: time="2026-03-07T01:49:36.087850887Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6bf85f8dd-c4kv7,Uid:856fa1da-0776-4eac-86d1-7571f0b78b29,Namespace:tigera-operator,Attempt:0,}" Mar 7 01:49:36.135650 containerd[1817]: time="2026-03-07T01:49:36.135408405Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:49:36.135650 containerd[1817]: time="2026-03-07T01:49:36.135472485Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:49:36.135650 containerd[1817]: time="2026-03-07T01:49:36.135488605Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:36.135650 containerd[1817]: time="2026-03-07T01:49:36.135568925Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:36.178215 containerd[1817]: time="2026-03-07T01:49:36.178175476Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6bf85f8dd-c4kv7,Uid:856fa1da-0776-4eac-86d1-7571f0b78b29,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"b75281193be8a818ed93f5c7aa6275cbc7b0db27e9980792b416c1f7868e285d\"" Mar 7 01:49:36.180500 containerd[1817]: time="2026-03-07T01:49:36.180315599Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\"" Mar 7 01:49:36.320367 containerd[1817]: time="2026-03-07T01:49:36.320325950Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8lqsm,Uid:6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb,Namespace:kube-system,Attempt:0,}" Mar 7 01:49:36.363889 containerd[1817]: time="2026-03-07T01:49:36.363695742Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:49:36.364025 containerd[1817]: time="2026-03-07T01:49:36.363744382Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:49:36.364519 containerd[1817]: time="2026-03-07T01:49:36.364332903Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:36.364519 containerd[1817]: time="2026-03-07T01:49:36.364463503Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:36.395973 containerd[1817]: time="2026-03-07T01:49:36.395882275Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8lqsm,Uid:6c4ffdc3-0e26-45d6-af9f-b529c6c6fddb,Namespace:kube-system,Attempt:0,} returns sandbox id \"3f48d92573861567dcf8759eb73a5ea1cc2579bd9cd2925a99605d43b7fa50da\"" Mar 7 01:49:36.406478 containerd[1817]: time="2026-03-07T01:49:36.406374972Z" level=info msg="CreateContainer within sandbox \"3f48d92573861567dcf8759eb73a5ea1cc2579bd9cd2925a99605d43b7fa50da\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 7 01:49:36.441584 containerd[1817]: time="2026-03-07T01:49:36.441525830Z" level=info msg="CreateContainer within sandbox \"3f48d92573861567dcf8759eb73a5ea1cc2579bd9cd2925a99605d43b7fa50da\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"adff417065bb960048a3c109f5119e757e00b58345739e0a4a197a0457e331c7\"" Mar 7 01:49:36.442375 containerd[1817]: time="2026-03-07T01:49:36.442342512Z" level=info msg="StartContainer for \"adff417065bb960048a3c109f5119e757e00b58345739e0a4a197a0457e331c7\"" Mar 7 01:49:36.498666 containerd[1817]: time="2026-03-07T01:49:36.498622165Z" level=info msg="StartContainer for \"adff417065bb960048a3c109f5119e757e00b58345739e0a4a197a0457e331c7\" returns successfully" Mar 7 01:49:36.872264 kubelet[3332]: I0307 01:49:36.871585 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-8lqsm" podStartSLOduration=1.87157294 podStartE2EDuration="1.87157294s" podCreationTimestamp="2026-03-07 01:49:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:49:36.607094264 +0000 UTC m=+8.154696300" watchObservedRunningTime="2026-03-07 01:49:36.87157294 +0000 UTC m=+8.419174976" Mar 7 01:49:37.983891 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount409453935.mount: Deactivated successfully. Mar 7 01:49:38.579222 containerd[1817]: time="2026-03-07T01:49:38.579169718Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:38.582397 containerd[1817]: time="2026-03-07T01:49:38.582233563Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.7: active requests=0, bytes read=25071565" Mar 7 01:49:38.585687 containerd[1817]: time="2026-03-07T01:49:38.585481489Z" level=info msg="ImageCreate event name:\"sha256:b2fef69c2456aa0a6f6dcb63425a69d11dc35a73b1883b250e4d92f5a697fefe\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:38.590199 containerd[1817]: time="2026-03-07T01:49:38.590151177Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:38.590967 containerd[1817]: time="2026-03-07T01:49:38.590936938Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.7\" with image id \"sha256:b2fef69c2456aa0a6f6dcb63425a69d11dc35a73b1883b250e4d92f5a697fefe\", repo tag \"quay.io/tigera/operator:v1.40.7\", repo digest \"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\", size \"25067560\" in 2.410584459s" Mar 7 01:49:38.591168 containerd[1817]: time="2026-03-07T01:49:38.590969458Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\" returns image reference \"sha256:b2fef69c2456aa0a6f6dcb63425a69d11dc35a73b1883b250e4d92f5a697fefe\"" Mar 7 01:49:38.599912 containerd[1817]: time="2026-03-07T01:49:38.599761792Z" level=info msg="CreateContainer within sandbox \"b75281193be8a818ed93f5c7aa6275cbc7b0db27e9980792b416c1f7868e285d\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 7 01:49:38.632053 containerd[1817]: time="2026-03-07T01:49:38.631947206Z" level=info msg="CreateContainer within sandbox \"b75281193be8a818ed93f5c7aa6275cbc7b0db27e9980792b416c1f7868e285d\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"e25538f9ce24f326c8b1a94e68b81c8a6da677d5313afdae8e8c44020d4edbb7\"" Mar 7 01:49:38.634856 containerd[1817]: time="2026-03-07T01:49:38.633696288Z" level=info msg="StartContainer for \"e25538f9ce24f326c8b1a94e68b81c8a6da677d5313afdae8e8c44020d4edbb7\"" Mar 7 01:49:38.678302 containerd[1817]: time="2026-03-07T01:49:38.678259242Z" level=info msg="StartContainer for \"e25538f9ce24f326c8b1a94e68b81c8a6da677d5313afdae8e8c44020d4edbb7\" returns successfully" Mar 7 01:49:39.760780 kubelet[3332]: I0307 01:49:39.760639 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6bf85f8dd-c4kv7" podStartSLOduration=2.348310606 podStartE2EDuration="4.760623788s" podCreationTimestamp="2026-03-07 01:49:35 +0000 UTC" firstStartedPulling="2026-03-07 01:49:36.179832118 +0000 UTC m=+7.727434154" lastFinishedPulling="2026-03-07 01:49:38.5921453 +0000 UTC m=+10.139747336" observedRunningTime="2026-03-07 01:49:39.619290275 +0000 UTC m=+11.166892311" watchObservedRunningTime="2026-03-07 01:49:39.760623788 +0000 UTC m=+11.308225824" Mar 7 01:49:44.571623 sudo[2341]: pam_unix(sudo:session): session closed for user root Mar 7 01:49:44.648811 sshd[2337]: pam_unix(sshd:session): session closed for user core Mar 7 01:49:44.656046 systemd-logind[1784]: Session 9 logged out. Waiting for processes to exit. Mar 7 01:49:44.658767 systemd[1]: sshd@6-10.200.20.15:22-10.200.16.10:33828.service: Deactivated successfully. Mar 7 01:49:44.666731 systemd[1]: session-9.scope: Deactivated successfully. Mar 7 01:49:44.669508 systemd-logind[1784]: Removed session 9. Mar 7 01:49:46.946349 kubelet[3332]: I0307 01:49:46.946293 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/1dc23b0d-ef5b-460e-9cfe-417f6a42be65-typha-certs\") pod \"calico-typha-9c79b7ddc-4lm27\" (UID: \"1dc23b0d-ef5b-460e-9cfe-417f6a42be65\") " pod="calico-system/calico-typha-9c79b7ddc-4lm27" Mar 7 01:49:46.946349 kubelet[3332]: I0307 01:49:46.946336 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd97h\" (UniqueName: \"kubernetes.io/projected/1dc23b0d-ef5b-460e-9cfe-417f6a42be65-kube-api-access-kd97h\") pod \"calico-typha-9c79b7ddc-4lm27\" (UID: \"1dc23b0d-ef5b-460e-9cfe-417f6a42be65\") " pod="calico-system/calico-typha-9c79b7ddc-4lm27" Mar 7 01:49:46.946349 kubelet[3332]: I0307 01:49:46.946357 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1dc23b0d-ef5b-460e-9cfe-417f6a42be65-tigera-ca-bundle\") pod \"calico-typha-9c79b7ddc-4lm27\" (UID: \"1dc23b0d-ef5b-460e-9cfe-417f6a42be65\") " pod="calico-system/calico-typha-9c79b7ddc-4lm27" Mar 7 01:49:47.047250 kubelet[3332]: I0307 01:49:47.047208 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq2mb\" (UniqueName: \"kubernetes.io/projected/72b38734-8960-422a-9a16-0b9e92c475bb-kube-api-access-jq2mb\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047250 kubelet[3332]: I0307 01:49:47.047259 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-lib-modules\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047398 kubelet[3332]: I0307 01:49:47.047279 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-nodeproc\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047398 kubelet[3332]: I0307 01:49:47.047293 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-policysync\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047398 kubelet[3332]: I0307 01:49:47.047307 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-var-lib-calico\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047398 kubelet[3332]: I0307 01:49:47.047320 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-xtables-lock\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047398 kubelet[3332]: I0307 01:49:47.047334 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-cni-log-dir\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047566 kubelet[3332]: I0307 01:49:47.047358 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-cni-bin-dir\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047566 kubelet[3332]: I0307 01:49:47.047371 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-sys-fs\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047566 kubelet[3332]: I0307 01:49:47.047385 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72b38734-8960-422a-9a16-0b9e92c475bb-tigera-ca-bundle\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047566 kubelet[3332]: I0307 01:49:47.047399 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-bpffs\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047566 kubelet[3332]: I0307 01:49:47.047414 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-flexvol-driver-host\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047699 kubelet[3332]: I0307 01:49:47.047428 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/72b38734-8960-422a-9a16-0b9e92c475bb-node-certs\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047699 kubelet[3332]: I0307 01:49:47.047452 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-var-run-calico\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.047699 kubelet[3332]: I0307 01:49:47.047472 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/72b38734-8960-422a-9a16-0b9e92c475bb-cni-net-dir\") pod \"calico-node-5whnz\" (UID: \"72b38734-8960-422a-9a16-0b9e92c475bb\") " pod="calico-system/calico-node-5whnz" Mar 7 01:49:47.138232 kubelet[3332]: E0307 01:49:47.138185 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:49:47.150343 kubelet[3332]: I0307 01:49:47.148033 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55v4l\" (UniqueName: \"kubernetes.io/projected/c189a211-99d4-4133-a1d8-fbd792f2c0a5-kube-api-access-55v4l\") pod \"csi-node-driver-cz2s4\" (UID: \"c189a211-99d4-4133-a1d8-fbd792f2c0a5\") " pod="calico-system/csi-node-driver-cz2s4" Mar 7 01:49:47.150343 kubelet[3332]: I0307 01:49:47.148078 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c189a211-99d4-4133-a1d8-fbd792f2c0a5-socket-dir\") pod \"csi-node-driver-cz2s4\" (UID: \"c189a211-99d4-4133-a1d8-fbd792f2c0a5\") " pod="calico-system/csi-node-driver-cz2s4" Mar 7 01:49:47.150343 kubelet[3332]: I0307 01:49:47.148137 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c189a211-99d4-4133-a1d8-fbd792f2c0a5-registration-dir\") pod \"csi-node-driver-cz2s4\" (UID: \"c189a211-99d4-4133-a1d8-fbd792f2c0a5\") " pod="calico-system/csi-node-driver-cz2s4" Mar 7 01:49:47.150343 kubelet[3332]: I0307 01:49:47.148177 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/c189a211-99d4-4133-a1d8-fbd792f2c0a5-varrun\") pod \"csi-node-driver-cz2s4\" (UID: \"c189a211-99d4-4133-a1d8-fbd792f2c0a5\") " pod="calico-system/csi-node-driver-cz2s4" Mar 7 01:49:47.150343 kubelet[3332]: I0307 01:49:47.148250 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c189a211-99d4-4133-a1d8-fbd792f2c0a5-kubelet-dir\") pod \"csi-node-driver-cz2s4\" (UID: \"c189a211-99d4-4133-a1d8-fbd792f2c0a5\") " pod="calico-system/csi-node-driver-cz2s4" Mar 7 01:49:47.152614 kubelet[3332]: E0307 01:49:47.152596 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.152705 kubelet[3332]: W0307 01:49:47.152692 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.152777 kubelet[3332]: E0307 01:49:47.152766 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.153049 kubelet[3332]: E0307 01:49:47.153035 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.153131 kubelet[3332]: W0307 01:49:47.153119 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.153187 kubelet[3332]: E0307 01:49:47.153177 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.153425 kubelet[3332]: E0307 01:49:47.153413 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.153834 kubelet[3332]: W0307 01:49:47.153816 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.153920 kubelet[3332]: E0307 01:49:47.153907 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.154924 kubelet[3332]: E0307 01:49:47.154907 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.155460 kubelet[3332]: W0307 01:49:47.155016 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.155584 kubelet[3332]: E0307 01:49:47.155570 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.156600 kubelet[3332]: E0307 01:49:47.156580 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.156695 kubelet[3332]: W0307 01:49:47.156682 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.157067 kubelet[3332]: E0307 01:49:47.157047 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.158171 kubelet[3332]: E0307 01:49:47.158150 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.158266 kubelet[3332]: W0307 01:49:47.158254 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.158321 kubelet[3332]: E0307 01:49:47.158311 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.159667 kubelet[3332]: E0307 01:49:47.159651 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.159745 kubelet[3332]: W0307 01:49:47.159733 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.159964 kubelet[3332]: E0307 01:49:47.159857 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.165023 kubelet[3332]: E0307 01:49:47.165005 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.168905 kubelet[3332]: W0307 01:49:47.167474 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.168905 kubelet[3332]: E0307 01:49:47.167505 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.171712 kubelet[3332]: E0307 01:49:47.171696 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.173473 kubelet[3332]: W0307 01:49:47.172033 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.174370 kubelet[3332]: E0307 01:49:47.173578 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.176569 kubelet[3332]: E0307 01:49:47.176550 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.176666 kubelet[3332]: W0307 01:49:47.176653 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.176839 kubelet[3332]: E0307 01:49:47.176714 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.177427 kubelet[3332]: E0307 01:49:47.177411 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.177615 kubelet[3332]: W0307 01:49:47.177497 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.177615 kubelet[3332]: E0307 01:49:47.177513 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.178831 kubelet[3332]: E0307 01:49:47.178814 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.179004 kubelet[3332]: W0307 01:49:47.178917 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.179004 kubelet[3332]: E0307 01:49:47.178935 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.179828 kubelet[3332]: E0307 01:49:47.179193 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.179828 kubelet[3332]: W0307 01:49:47.179205 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.179828 kubelet[3332]: E0307 01:49:47.179215 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.182115 kubelet[3332]: E0307 01:49:47.180598 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.182115 kubelet[3332]: W0307 01:49:47.180614 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.182115 kubelet[3332]: E0307 01:49:47.180626 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.191180 kubelet[3332]: E0307 01:49:47.190972 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.191180 kubelet[3332]: W0307 01:49:47.191002 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.191180 kubelet[3332]: E0307 01:49:47.191019 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.235458 containerd[1817]: time="2026-03-07T01:49:47.235276437Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-9c79b7ddc-4lm27,Uid:1dc23b0d-ef5b-460e-9cfe-417f6a42be65,Namespace:calico-system,Attempt:0,}" Mar 7 01:49:47.249870 kubelet[3332]: E0307 01:49:47.249138 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.249870 kubelet[3332]: W0307 01:49:47.249156 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.249870 kubelet[3332]: E0307 01:49:47.249173 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.251848 kubelet[3332]: E0307 01:49:47.250854 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.251848 kubelet[3332]: W0307 01:49:47.250867 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.251848 kubelet[3332]: E0307 01:49:47.250891 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.251848 kubelet[3332]: E0307 01:49:47.251128 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.251848 kubelet[3332]: W0307 01:49:47.251138 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.251848 kubelet[3332]: E0307 01:49:47.251149 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.251848 kubelet[3332]: E0307 01:49:47.251333 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.251848 kubelet[3332]: W0307 01:49:47.251342 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.251848 kubelet[3332]: E0307 01:49:47.251353 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.252746 kubelet[3332]: E0307 01:49:47.252316 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.252746 kubelet[3332]: W0307 01:49:47.252329 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.252746 kubelet[3332]: E0307 01:49:47.252341 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.253413 kubelet[3332]: E0307 01:49:47.252920 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.253413 kubelet[3332]: W0307 01:49:47.252937 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.253413 kubelet[3332]: E0307 01:49:47.252952 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.253770 kubelet[3332]: E0307 01:49:47.253631 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.253770 kubelet[3332]: W0307 01:49:47.253644 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.253770 kubelet[3332]: E0307 01:49:47.253656 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.254542 kubelet[3332]: E0307 01:49:47.254356 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.254542 kubelet[3332]: W0307 01:49:47.254370 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.254772 kubelet[3332]: E0307 01:49:47.254548 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.255634 kubelet[3332]: E0307 01:49:47.255613 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.255634 kubelet[3332]: W0307 01:49:47.255629 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.256503 kubelet[3332]: E0307 01:49:47.256472 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.256770 kubelet[3332]: E0307 01:49:47.256685 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.256770 kubelet[3332]: W0307 01:49:47.256768 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.257066 kubelet[3332]: E0307 01:49:47.256783 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.257066 kubelet[3332]: E0307 01:49:47.257048 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.257066 kubelet[3332]: W0307 01:49:47.257059 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.257136 kubelet[3332]: E0307 01:49:47.257078 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.257645 kubelet[3332]: E0307 01:49:47.257628 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.257645 kubelet[3332]: W0307 01:49:47.257641 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.258097 kubelet[3332]: E0307 01:49:47.257653 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.258097 kubelet[3332]: E0307 01:49:47.258007 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.258097 kubelet[3332]: W0307 01:49:47.258018 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.258097 kubelet[3332]: E0307 01:49:47.258028 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.258763 kubelet[3332]: E0307 01:49:47.258745 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.258763 kubelet[3332]: W0307 01:49:47.258760 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.259824 kubelet[3332]: E0307 01:49:47.258773 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.259824 kubelet[3332]: E0307 01:49:47.258954 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.259824 kubelet[3332]: W0307 01:49:47.258963 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.259824 kubelet[3332]: E0307 01:49:47.258972 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.260320 kubelet[3332]: E0307 01:49:47.260020 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.260320 kubelet[3332]: W0307 01:49:47.260034 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.260320 kubelet[3332]: E0307 01:49:47.260047 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.260692 kubelet[3332]: E0307 01:49:47.260353 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.260692 kubelet[3332]: W0307 01:49:47.260364 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.260692 kubelet[3332]: E0307 01:49:47.260375 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.261309 kubelet[3332]: E0307 01:49:47.260717 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.261309 kubelet[3332]: W0307 01:49:47.260728 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.261309 kubelet[3332]: E0307 01:49:47.260740 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.261826 kubelet[3332]: E0307 01:49:47.261810 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.261826 kubelet[3332]: W0307 01:49:47.261824 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.262354 kubelet[3332]: E0307 01:49:47.261836 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.262354 kubelet[3332]: E0307 01:49:47.262115 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.262354 kubelet[3332]: W0307 01:49:47.262128 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.262354 kubelet[3332]: E0307 01:49:47.262140 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.263416 kubelet[3332]: E0307 01:49:47.262611 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.263416 kubelet[3332]: W0307 01:49:47.262622 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.263416 kubelet[3332]: E0307 01:49:47.262634 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.263416 kubelet[3332]: E0307 01:49:47.263395 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.263416 kubelet[3332]: W0307 01:49:47.263407 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.263416 kubelet[3332]: E0307 01:49:47.263418 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.266599 kubelet[3332]: E0307 01:49:47.266499 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.266599 kubelet[3332]: W0307 01:49:47.266513 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.266599 kubelet[3332]: E0307 01:49:47.266526 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.268472 kubelet[3332]: E0307 01:49:47.267644 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.268472 kubelet[3332]: W0307 01:49:47.267658 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.268472 kubelet[3332]: E0307 01:49:47.267671 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.268717 kubelet[3332]: E0307 01:49:47.268700 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.268717 kubelet[3332]: W0307 01:49:47.268714 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.268779 kubelet[3332]: E0307 01:49:47.268726 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.286831 containerd[1817]: time="2026-03-07T01:49:47.286755963Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:49:47.286980 containerd[1817]: time="2026-03-07T01:49:47.286843843Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:49:47.286980 containerd[1817]: time="2026-03-07T01:49:47.286871083Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:47.287040 containerd[1817]: time="2026-03-07T01:49:47.286969323Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:47.289024 kubelet[3332]: E0307 01:49:47.288901 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:47.289024 kubelet[3332]: W0307 01:49:47.288916 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:47.289024 kubelet[3332]: E0307 01:49:47.288929 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:47.325178 containerd[1817]: time="2026-03-07T01:49:47.325070347Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-5whnz,Uid:72b38734-8960-422a-9a16-0b9e92c475bb,Namespace:calico-system,Attempt:0,}" Mar 7 01:49:47.334153 containerd[1817]: time="2026-03-07T01:49:47.334119042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-9c79b7ddc-4lm27,Uid:1dc23b0d-ef5b-460e-9cfe-417f6a42be65,Namespace:calico-system,Attempt:0,} returns sandbox id \"a702577719a82d17309c35bd809065ad55312441a0f4c229d028b2dd43263c91\"" Mar 7 01:49:47.335723 containerd[1817]: time="2026-03-07T01:49:47.335655565Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\"" Mar 7 01:49:47.367278 containerd[1817]: time="2026-03-07T01:49:47.367173738Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:49:47.367278 containerd[1817]: time="2026-03-07T01:49:47.367236818Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:49:47.367278 containerd[1817]: time="2026-03-07T01:49:47.367252738Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:47.367516 containerd[1817]: time="2026-03-07T01:49:47.367329258Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:49:47.403562 containerd[1817]: time="2026-03-07T01:49:47.403523598Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-5whnz,Uid:72b38734-8960-422a-9a16-0b9e92c475bb,Namespace:calico-system,Attempt:0,} returns sandbox id \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\"" Mar 7 01:49:48.695268 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1797968498.mount: Deactivated successfully. Mar 7 01:49:49.547715 kubelet[3332]: E0307 01:49:49.547667 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:49:49.813481 containerd[1817]: time="2026-03-07T01:49:49.812618890Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:49.817350 containerd[1817]: time="2026-03-07T01:49:49.817320139Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.4: active requests=0, bytes read=33865174" Mar 7 01:49:49.820590 containerd[1817]: time="2026-03-07T01:49:49.820532345Z" level=info msg="ImageCreate event name:\"sha256:e836e1dea560d4c477b347f1c93c245aec618361306b23eda1d6bb7665476182\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:49.825594 containerd[1817]: time="2026-03-07T01:49:49.825549914Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:49.826463 containerd[1817]: time="2026-03-07T01:49:49.826340115Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.4\" with image id \"sha256:e836e1dea560d4c477b347f1c93c245aec618361306b23eda1d6bb7665476182\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\", size \"33865028\" in 2.49061555s" Mar 7 01:49:49.826463 containerd[1817]: time="2026-03-07T01:49:49.826371475Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\" returns image reference \"sha256:e836e1dea560d4c477b347f1c93c245aec618361306b23eda1d6bb7665476182\"" Mar 7 01:49:49.827949 containerd[1817]: time="2026-03-07T01:49:49.827790038Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\"" Mar 7 01:49:49.846698 containerd[1817]: time="2026-03-07T01:49:49.846664952Z" level=info msg="CreateContainer within sandbox \"a702577719a82d17309c35bd809065ad55312441a0f4c229d028b2dd43263c91\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 7 01:49:49.882274 containerd[1817]: time="2026-03-07T01:49:49.882174936Z" level=info msg="CreateContainer within sandbox \"a702577719a82d17309c35bd809065ad55312441a0f4c229d028b2dd43263c91\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"bee6b87c0e23a00f2e3912e030a94a20dfa15d44adc7f884387df3ea24142365\"" Mar 7 01:49:49.883081 containerd[1817]: time="2026-03-07T01:49:49.882713657Z" level=info msg="StartContainer for \"bee6b87c0e23a00f2e3912e030a94a20dfa15d44adc7f884387df3ea24142365\"" Mar 7 01:49:49.943151 containerd[1817]: time="2026-03-07T01:49:49.943036565Z" level=info msg="StartContainer for \"bee6b87c0e23a00f2e3912e030a94a20dfa15d44adc7f884387df3ea24142365\" returns successfully" Mar 7 01:49:50.643377 kubelet[3332]: I0307 01:49:50.643316 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-9c79b7ddc-4lm27" podStartSLOduration=2.151226353 podStartE2EDuration="4.643297346s" podCreationTimestamp="2026-03-07 01:49:46 +0000 UTC" firstStartedPulling="2026-03-07 01:49:47.335207564 +0000 UTC m=+18.882809600" lastFinishedPulling="2026-03-07 01:49:49.827278597 +0000 UTC m=+21.374880593" observedRunningTime="2026-03-07 01:49:50.642547264 +0000 UTC m=+22.190149300" watchObservedRunningTime="2026-03-07 01:49:50.643297346 +0000 UTC m=+22.190899422" Mar 7 01:49:50.661877 kubelet[3332]: E0307 01:49:50.661847 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.661877 kubelet[3332]: W0307 01:49:50.661871 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.662383 kubelet[3332]: E0307 01:49:50.661890 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.662535 kubelet[3332]: E0307 01:49:50.662514 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.662568 kubelet[3332]: W0307 01:49:50.662526 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.662871 kubelet[3332]: E0307 01:49:50.662567 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.662871 kubelet[3332]: E0307 01:49:50.662724 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.662871 kubelet[3332]: W0307 01:49:50.662732 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.662871 kubelet[3332]: E0307 01:49:50.662758 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.663088 kubelet[3332]: E0307 01:49:50.663062 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.663088 kubelet[3332]: W0307 01:49:50.663077 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.663145 kubelet[3332]: E0307 01:49:50.663089 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.663272 kubelet[3332]: E0307 01:49:50.663256 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.663272 kubelet[3332]: W0307 01:49:50.663270 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.663329 kubelet[3332]: E0307 01:49:50.663280 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.663425 kubelet[3332]: E0307 01:49:50.663410 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.663425 kubelet[3332]: W0307 01:49:50.663421 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.663495 kubelet[3332]: E0307 01:49:50.663432 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.663602 kubelet[3332]: E0307 01:49:50.663584 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.663602 kubelet[3332]: W0307 01:49:50.663596 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.663653 kubelet[3332]: E0307 01:49:50.663604 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.663759 kubelet[3332]: E0307 01:49:50.663741 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.663759 kubelet[3332]: W0307 01:49:50.663752 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.663812 kubelet[3332]: E0307 01:49:50.663760 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.663925 kubelet[3332]: E0307 01:49:50.663901 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.663925 kubelet[3332]: W0307 01:49:50.663912 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.663925 kubelet[3332]: E0307 01:49:50.663921 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.664060 kubelet[3332]: E0307 01:49:50.664048 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.664060 kubelet[3332]: W0307 01:49:50.664058 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.664128 kubelet[3332]: E0307 01:49:50.664066 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.664206 kubelet[3332]: E0307 01:49:50.664192 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.664206 kubelet[3332]: W0307 01:49:50.664203 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.664279 kubelet[3332]: E0307 01:49:50.664212 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.664507 kubelet[3332]: E0307 01:49:50.664465 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.664507 kubelet[3332]: W0307 01:49:50.664499 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.664914 kubelet[3332]: E0307 01:49:50.664509 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.665237 kubelet[3332]: E0307 01:49:50.665211 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.665237 kubelet[3332]: W0307 01:49:50.665232 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.665322 kubelet[3332]: E0307 01:49:50.665244 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.666344 kubelet[3332]: E0307 01:49:50.666266 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.666344 kubelet[3332]: W0307 01:49:50.666279 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.666581 kubelet[3332]: E0307 01:49:50.666387 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.666581 kubelet[3332]: E0307 01:49:50.666568 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.666581 kubelet[3332]: W0307 01:49:50.666578 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.666771 kubelet[3332]: E0307 01:49:50.666588 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.680801 kubelet[3332]: E0307 01:49:50.680782 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.680801 kubelet[3332]: W0307 01:49:50.680797 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.681099 kubelet[3332]: E0307 01:49:50.680810 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.681099 kubelet[3332]: E0307 01:49:50.680988 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.681099 kubelet[3332]: W0307 01:49:50.680996 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.681099 kubelet[3332]: E0307 01:49:50.681005 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.682732 kubelet[3332]: E0307 01:49:50.682714 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.682732 kubelet[3332]: W0307 01:49:50.682729 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.682910 kubelet[3332]: E0307 01:49:50.682741 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.682910 kubelet[3332]: E0307 01:49:50.682906 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.682988 kubelet[3332]: W0307 01:49:50.682914 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.682988 kubelet[3332]: E0307 01:49:50.682923 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.683100 kubelet[3332]: E0307 01:49:50.683087 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.683100 kubelet[3332]: W0307 01:49:50.683098 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.683175 kubelet[3332]: E0307 01:49:50.683107 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.683252 kubelet[3332]: E0307 01:49:50.683239 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.683252 kubelet[3332]: W0307 01:49:50.683247 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.683501 kubelet[3332]: E0307 01:49:50.683256 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.683501 kubelet[3332]: E0307 01:49:50.683390 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.683501 kubelet[3332]: W0307 01:49:50.683397 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.683501 kubelet[3332]: E0307 01:49:50.683405 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.685221 kubelet[3332]: E0307 01:49:50.685190 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.685221 kubelet[3332]: W0307 01:49:50.685202 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.685221 kubelet[3332]: E0307 01:49:50.685213 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.685587 kubelet[3332]: E0307 01:49:50.685570 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.685587 kubelet[3332]: W0307 01:49:50.685584 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.685673 kubelet[3332]: E0307 01:49:50.685595 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.685743 kubelet[3332]: E0307 01:49:50.685734 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.685743 kubelet[3332]: W0307 01:49:50.685742 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.685743 kubelet[3332]: E0307 01:49:50.685750 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.685888 kubelet[3332]: E0307 01:49:50.685875 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.685888 kubelet[3332]: W0307 01:49:50.685885 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.685971 kubelet[3332]: E0307 01:49:50.685893 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.686202 kubelet[3332]: E0307 01:49:50.686024 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.686202 kubelet[3332]: W0307 01:49:50.686031 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.686202 kubelet[3332]: E0307 01:49:50.686039 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.686589 kubelet[3332]: E0307 01:49:50.686355 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.686589 kubelet[3332]: W0307 01:49:50.686369 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.686589 kubelet[3332]: E0307 01:49:50.686379 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.687115 kubelet[3332]: E0307 01:49:50.686657 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.687115 kubelet[3332]: W0307 01:49:50.686666 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.687115 kubelet[3332]: E0307 01:49:50.686678 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.688179 kubelet[3332]: E0307 01:49:50.688158 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.688179 kubelet[3332]: W0307 01:49:50.688174 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.688933 kubelet[3332]: E0307 01:49:50.688187 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.688933 kubelet[3332]: E0307 01:49:50.688725 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.688933 kubelet[3332]: W0307 01:49:50.688736 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.688933 kubelet[3332]: E0307 01:49:50.688746 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.689051 kubelet[3332]: E0307 01:49:50.688982 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.689051 kubelet[3332]: W0307 01:49:50.688992 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.689051 kubelet[3332]: E0307 01:49:50.689003 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:50.689352 kubelet[3332]: E0307 01:49:50.689336 3332 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 7 01:49:50.689352 kubelet[3332]: W0307 01:49:50.689348 3332 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 7 01:49:50.689352 kubelet[3332]: E0307 01:49:50.689359 3332 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 7 01:49:51.450383 containerd[1817]: time="2026-03-07T01:49:51.450329359Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:51.453754 containerd[1817]: time="2026-03-07T01:49:51.453559124Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4: active requests=0, bytes read=4457682" Mar 7 01:49:51.458476 containerd[1817]: time="2026-03-07T01:49:51.457510932Z" level=info msg="ImageCreate event name:\"sha256:449a6463eaa02e13b190ef7c4057191febcc65ab9418bae3bc0995f5bce65798\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:51.463004 containerd[1817]: time="2026-03-07T01:49:51.462459380Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:51.464384 containerd[1817]: time="2026-03-07T01:49:51.464346584Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" with image id \"sha256:449a6463eaa02e13b190ef7c4057191febcc65ab9418bae3bc0995f5bce65798\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\", size \"5855167\" in 1.636526426s" Mar 7 01:49:51.464384 containerd[1817]: time="2026-03-07T01:49:51.464383664Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" returns image reference \"sha256:449a6463eaa02e13b190ef7c4057191febcc65ab9418bae3bc0995f5bce65798\"" Mar 7 01:49:51.471308 containerd[1817]: time="2026-03-07T01:49:51.471274116Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 7 01:49:51.509636 containerd[1817]: time="2026-03-07T01:49:51.509578985Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"0785afdd684f46ad3f771915af4f4f83d8d6f9c2c86f2082e63c965c47ab809c\"" Mar 7 01:49:51.510393 containerd[1817]: time="2026-03-07T01:49:51.510199306Z" level=info msg="StartContainer for \"0785afdd684f46ad3f771915af4f4f83d8d6f9c2c86f2082e63c965c47ab809c\"" Mar 7 01:49:51.537576 systemd[1]: run-containerd-runc-k8s.io-0785afdd684f46ad3f771915af4f4f83d8d6f9c2c86f2082e63c965c47ab809c-runc.DhtGzQ.mount: Deactivated successfully. Mar 7 01:49:51.548343 kubelet[3332]: E0307 01:49:51.548302 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:49:51.568610 containerd[1817]: time="2026-03-07T01:49:51.568520691Z" level=info msg="StartContainer for \"0785afdd684f46ad3f771915af4f4f83d8d6f9c2c86f2082e63c965c47ab809c\" returns successfully" Mar 7 01:49:52.492972 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0785afdd684f46ad3f771915af4f4f83d8d6f9c2c86f2082e63c965c47ab809c-rootfs.mount: Deactivated successfully. Mar 7 01:49:52.711476 containerd[1817]: time="2026-03-07T01:49:52.711311549Z" level=info msg="shim disconnected" id=0785afdd684f46ad3f771915af4f4f83d8d6f9c2c86f2082e63c965c47ab809c namespace=k8s.io Mar 7 01:49:52.711476 containerd[1817]: time="2026-03-07T01:49:52.711366749Z" level=warning msg="cleaning up after shim disconnected" id=0785afdd684f46ad3f771915af4f4f83d8d6f9c2c86f2082e63c965c47ab809c namespace=k8s.io Mar 7 01:49:52.711476 containerd[1817]: time="2026-03-07T01:49:52.711375509Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:49:53.547793 kubelet[3332]: E0307 01:49:53.547740 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:49:53.637214 containerd[1817]: time="2026-03-07T01:49:53.637084615Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\"" Mar 7 01:49:55.548034 kubelet[3332]: E0307 01:49:55.547982 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:49:57.548244 kubelet[3332]: E0307 01:49:57.547893 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:49:59.548084 kubelet[3332]: E0307 01:49:59.547986 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:49:59.666037 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2595093340.mount: Deactivated successfully. Mar 7 01:49:59.713065 containerd[1817]: time="2026-03-07T01:49:59.713012527Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:59.718786 containerd[1817]: time="2026-03-07T01:49:59.718670537Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.4: active requests=0, bytes read=153921674" Mar 7 01:49:59.722199 containerd[1817]: time="2026-03-07T01:49:59.721976343Z" level=info msg="ImageCreate event name:\"sha256:27be54f2b9e47d96c7e9e5ad16e26ec298c1829f31885c81a622d50472c8ac97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:59.726652 containerd[1817]: time="2026-03-07T01:49:59.726582232Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:49:59.727708 containerd[1817]: time="2026-03-07T01:49:59.727209033Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.4\" with image id \"sha256:27be54f2b9e47d96c7e9e5ad16e26ec298c1829f31885c81a622d50472c8ac97\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\", size \"153921536\" in 6.090087378s" Mar 7 01:49:59.727708 containerd[1817]: time="2026-03-07T01:49:59.727248713Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\" returns image reference \"sha256:27be54f2b9e47d96c7e9e5ad16e26ec298c1829f31885c81a622d50472c8ac97\"" Mar 7 01:49:59.734646 containerd[1817]: time="2026-03-07T01:49:59.734614927Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Mar 7 01:49:59.774417 containerd[1817]: time="2026-03-07T01:49:59.774377361Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"4f28157d14a69c03489e8ebc2d1f2af94022207e46c3e9e9760ed7a659593ca0\"" Mar 7 01:49:59.775286 containerd[1817]: time="2026-03-07T01:49:59.775264562Z" level=info msg="StartContainer for \"4f28157d14a69c03489e8ebc2d1f2af94022207e46c3e9e9760ed7a659593ca0\"" Mar 7 01:49:59.840418 containerd[1817]: time="2026-03-07T01:49:59.840144323Z" level=info msg="StartContainer for \"4f28157d14a69c03489e8ebc2d1f2af94022207e46c3e9e9760ed7a659593ca0\" returns successfully" Mar 7 01:50:00.663979 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4f28157d14a69c03489e8ebc2d1f2af94022207e46c3e9e9760ed7a659593ca0-rootfs.mount: Deactivated successfully. Mar 7 01:50:01.547725 kubelet[3332]: E0307 01:50:01.547679 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:50:01.550694 containerd[1817]: time="2026-03-07T01:50:01.550637261Z" level=info msg="shim disconnected" id=4f28157d14a69c03489e8ebc2d1f2af94022207e46c3e9e9760ed7a659593ca0 namespace=k8s.io Mar 7 01:50:01.550694 containerd[1817]: time="2026-03-07T01:50:01.550692421Z" level=warning msg="cleaning up after shim disconnected" id=4f28157d14a69c03489e8ebc2d1f2af94022207e46c3e9e9760ed7a659593ca0 namespace=k8s.io Mar 7 01:50:01.550994 containerd[1817]: time="2026-03-07T01:50:01.550707941Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:50:01.654455 containerd[1817]: time="2026-03-07T01:50:01.654373934Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\"" Mar 7 01:50:03.549131 kubelet[3332]: E0307 01:50:03.548168 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:50:05.431741 containerd[1817]: time="2026-03-07T01:50:05.431687156Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:05.434976 containerd[1817]: time="2026-03-07T01:50:05.434947402Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.4: active requests=0, bytes read=66009216" Mar 7 01:50:05.438675 containerd[1817]: time="2026-03-07T01:50:05.438646689Z" level=info msg="ImageCreate event name:\"sha256:c10bed152367fad8c19e9400f12b748d6fbc20498086983df13e70e36f24511b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:05.443215 containerd[1817]: time="2026-03-07T01:50:05.443182497Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:05.444297 containerd[1817]: time="2026-03-07T01:50:05.443795818Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.4\" with image id \"sha256:c10bed152367fad8c19e9400f12b748d6fbc20498086983df13e70e36f24511b\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\", size \"67406741\" in 3.789381444s" Mar 7 01:50:05.444297 containerd[1817]: time="2026-03-07T01:50:05.443827979Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\" returns image reference \"sha256:c10bed152367fad8c19e9400f12b748d6fbc20498086983df13e70e36f24511b\"" Mar 7 01:50:05.456046 containerd[1817]: time="2026-03-07T01:50:05.456014641Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 7 01:50:05.493829 containerd[1817]: time="2026-03-07T01:50:05.493783792Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"55d4eda490a69e340605e05ef984ff3a1c42d869b71430359a1f40b39c2146f2\"" Mar 7 01:50:05.494417 containerd[1817]: time="2026-03-07T01:50:05.494359193Z" level=info msg="StartContainer for \"55d4eda490a69e340605e05ef984ff3a1c42d869b71430359a1f40b39c2146f2\"" Mar 7 01:50:05.547563 kubelet[3332]: E0307 01:50:05.547513 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:50:05.549413 containerd[1817]: time="2026-03-07T01:50:05.549377535Z" level=info msg="StartContainer for \"55d4eda490a69e340605e05ef984ff3a1c42d869b71430359a1f40b39c2146f2\" returns successfully" Mar 7 01:50:07.548785 kubelet[3332]: E0307 01:50:07.547783 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-cz2s4" podUID="c189a211-99d4-4133-a1d8-fbd792f2c0a5" Mar 7 01:50:07.774297 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-55d4eda490a69e340605e05ef984ff3a1c42d869b71430359a1f40b39c2146f2-rootfs.mount: Deactivated successfully. Mar 7 01:50:07.778030 kubelet[3332]: I0307 01:50:07.775157 3332 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Mar 7 01:50:07.786193 containerd[1817]: time="2026-03-07T01:50:07.784360539Z" level=info msg="shim disconnected" id=55d4eda490a69e340605e05ef984ff3a1c42d869b71430359a1f40b39c2146f2 namespace=k8s.io Mar 7 01:50:07.786193 containerd[1817]: time="2026-03-07T01:50:07.785188220Z" level=warning msg="cleaning up after shim disconnected" id=55d4eda490a69e340605e05ef984ff3a1c42d869b71430359a1f40b39c2146f2 namespace=k8s.io Mar 7 01:50:07.786193 containerd[1817]: time="2026-03-07T01:50:07.785203300Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 01:50:07.894337 kubelet[3332]: I0307 01:50:07.894299 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k28r\" (UniqueName: \"kubernetes.io/projected/f64343b8-a2cd-4274-908a-c1805396f8d4-kube-api-access-9k28r\") pod \"calico-apiserver-57bc98fcf9-gb6nc\" (UID: \"f64343b8-a2cd-4274-908a-c1805396f8d4\") " pod="calico-system/calico-apiserver-57bc98fcf9-gb6nc" Mar 7 01:50:07.894337 kubelet[3332]: I0307 01:50:07.894342 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a40182e6-4b51-42f9-8f1a-ad85b1365603-config-volume\") pod \"coredns-674b8bbfcf-hkgbg\" (UID: \"a40182e6-4b51-42f9-8f1a-ad85b1365603\") " pod="kube-system/coredns-674b8bbfcf-hkgbg" Mar 7 01:50:07.894525 kubelet[3332]: I0307 01:50:07.894363 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkt9q\" (UniqueName: \"kubernetes.io/projected/da70d8fb-e941-4544-a89f-11fb7be7f164-kube-api-access-lkt9q\") pod \"whisker-6b776dffb4-p5tkl\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " pod="calico-system/whisker-6b776dffb4-p5tkl" Mar 7 01:50:07.894525 kubelet[3332]: I0307 01:50:07.894383 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cd6b288-aa15-4bc8-b027-e2e3c62e23a5-config-volume\") pod \"coredns-674b8bbfcf-n7zbg\" (UID: \"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5\") " pod="kube-system/coredns-674b8bbfcf-n7zbg" Mar 7 01:50:07.894525 kubelet[3332]: I0307 01:50:07.894403 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4wlw\" (UniqueName: \"kubernetes.io/projected/a40182e6-4b51-42f9-8f1a-ad85b1365603-kube-api-access-h4wlw\") pod \"coredns-674b8bbfcf-hkgbg\" (UID: \"a40182e6-4b51-42f9-8f1a-ad85b1365603\") " pod="kube-system/coredns-674b8bbfcf-hkgbg" Mar 7 01:50:07.894525 kubelet[3332]: I0307 01:50:07.894422 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-backend-key-pair\") pod \"whisker-6b776dffb4-p5tkl\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " pod="calico-system/whisker-6b776dffb4-p5tkl" Mar 7 01:50:07.894525 kubelet[3332]: I0307 01:50:07.894457 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f64343b8-a2cd-4274-908a-c1805396f8d4-calico-apiserver-certs\") pod \"calico-apiserver-57bc98fcf9-gb6nc\" (UID: \"f64343b8-a2cd-4274-908a-c1805396f8d4\") " pod="calico-system/calico-apiserver-57bc98fcf9-gb6nc" Mar 7 01:50:07.894645 kubelet[3332]: I0307 01:50:07.894477 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv2cj\" (UniqueName: \"kubernetes.io/projected/58c58ebb-a76c-40e3-bec1-27853dcf0611-kube-api-access-cv2cj\") pod \"calico-apiserver-57bc98fcf9-qfvhc\" (UID: \"58c58ebb-a76c-40e3-bec1-27853dcf0611\") " pod="calico-system/calico-apiserver-57bc98fcf9-qfvhc" Mar 7 01:50:07.894645 kubelet[3332]: I0307 01:50:07.894496 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8de81ec4-1390-4c8d-8e83-a56e0632cd6d-tigera-ca-bundle\") pod \"calico-kube-controllers-68c448d44-wz8jw\" (UID: \"8de81ec4-1390-4c8d-8e83-a56e0632cd6d\") " pod="calico-system/calico-kube-controllers-68c448d44-wz8jw" Mar 7 01:50:07.894645 kubelet[3332]: I0307 01:50:07.894512 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqzp5\" (UniqueName: \"kubernetes.io/projected/8cd6b288-aa15-4bc8-b027-e2e3c62e23a5-kube-api-access-gqzp5\") pod \"coredns-674b8bbfcf-n7zbg\" (UID: \"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5\") " pod="kube-system/coredns-674b8bbfcf-n7zbg" Mar 7 01:50:07.894645 kubelet[3332]: I0307 01:50:07.894527 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-nginx-config\") pod \"whisker-6b776dffb4-p5tkl\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " pod="calico-system/whisker-6b776dffb4-p5tkl" Mar 7 01:50:07.894645 kubelet[3332]: I0307 01:50:07.894542 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-ca-bundle\") pod \"whisker-6b776dffb4-p5tkl\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " pod="calico-system/whisker-6b776dffb4-p5tkl" Mar 7 01:50:07.894755 kubelet[3332]: I0307 01:50:07.894560 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqsrx\" (UniqueName: \"kubernetes.io/projected/e48198d3-aa71-41df-a131-1b93c5c926c5-kube-api-access-tqsrx\") pod \"goldmane-5b85766d88-58627\" (UID: \"e48198d3-aa71-41df-a131-1b93c5c926c5\") " pod="calico-system/goldmane-5b85766d88-58627" Mar 7 01:50:07.894755 kubelet[3332]: I0307 01:50:07.894577 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e48198d3-aa71-41df-a131-1b93c5c926c5-config\") pod \"goldmane-5b85766d88-58627\" (UID: \"e48198d3-aa71-41df-a131-1b93c5c926c5\") " pod="calico-system/goldmane-5b85766d88-58627" Mar 7 01:50:07.894755 kubelet[3332]: I0307 01:50:07.894591 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/e48198d3-aa71-41df-a131-1b93c5c926c5-goldmane-key-pair\") pod \"goldmane-5b85766d88-58627\" (UID: \"e48198d3-aa71-41df-a131-1b93c5c926c5\") " pod="calico-system/goldmane-5b85766d88-58627" Mar 7 01:50:07.894755 kubelet[3332]: I0307 01:50:07.894609 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg2vm\" (UniqueName: \"kubernetes.io/projected/8de81ec4-1390-4c8d-8e83-a56e0632cd6d-kube-api-access-gg2vm\") pod \"calico-kube-controllers-68c448d44-wz8jw\" (UID: \"8de81ec4-1390-4c8d-8e83-a56e0632cd6d\") " pod="calico-system/calico-kube-controllers-68c448d44-wz8jw" Mar 7 01:50:07.894755 kubelet[3332]: I0307 01:50:07.894629 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e48198d3-aa71-41df-a131-1b93c5c926c5-goldmane-ca-bundle\") pod \"goldmane-5b85766d88-58627\" (UID: \"e48198d3-aa71-41df-a131-1b93c5c926c5\") " pod="calico-system/goldmane-5b85766d88-58627" Mar 7 01:50:07.894863 kubelet[3332]: I0307 01:50:07.894643 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/58c58ebb-a76c-40e3-bec1-27853dcf0611-calico-apiserver-certs\") pod \"calico-apiserver-57bc98fcf9-qfvhc\" (UID: \"58c58ebb-a76c-40e3-bec1-27853dcf0611\") " pod="calico-system/calico-apiserver-57bc98fcf9-qfvhc" Mar 7 01:50:08.129983 containerd[1817]: time="2026-03-07T01:50:08.129938503Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-58627,Uid:e48198d3-aa71-41df-a131-1b93c5c926c5,Namespace:calico-system,Attempt:0,}" Mar 7 01:50:08.138954 containerd[1817]: time="2026-03-07T01:50:08.138926519Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n7zbg,Uid:8cd6b288-aa15-4bc8-b027-e2e3c62e23a5,Namespace:kube-system,Attempt:0,}" Mar 7 01:50:08.145753 containerd[1817]: time="2026-03-07T01:50:08.145669372Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hkgbg,Uid:a40182e6-4b51-42f9-8f1a-ad85b1365603,Namespace:kube-system,Attempt:0,}" Mar 7 01:50:08.149408 containerd[1817]: time="2026-03-07T01:50:08.149378819Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-68c448d44-wz8jw,Uid:8de81ec4-1390-4c8d-8e83-a56e0632cd6d,Namespace:calico-system,Attempt:0,}" Mar 7 01:50:08.156242 containerd[1817]: time="2026-03-07T01:50:08.156137632Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6b776dffb4-p5tkl,Uid:da70d8fb-e941-4544-a89f-11fb7be7f164,Namespace:calico-system,Attempt:0,}" Mar 7 01:50:08.162004 containerd[1817]: time="2026-03-07T01:50:08.161827522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-qfvhc,Uid:58c58ebb-a76c-40e3-bec1-27853dcf0611,Namespace:calico-system,Attempt:0,}" Mar 7 01:50:08.166002 containerd[1817]: time="2026-03-07T01:50:08.165587969Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-gb6nc,Uid:f64343b8-a2cd-4274-908a-c1805396f8d4,Namespace:calico-system,Attempt:0,}" Mar 7 01:50:08.481938 containerd[1817]: time="2026-03-07T01:50:08.481808878Z" level=error msg="Failed to destroy network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.482583 containerd[1817]: time="2026-03-07T01:50:08.482552360Z" level=error msg="encountered an error cleaning up failed sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.482721 containerd[1817]: time="2026-03-07T01:50:08.482698880Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-58627,Uid:e48198d3-aa71-41df-a131-1b93c5c926c5,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.483065 kubelet[3332]: E0307 01:50:08.483025 3332 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.485334 kubelet[3332]: E0307 01:50:08.483211 3332 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-5b85766d88-58627" Mar 7 01:50:08.485334 kubelet[3332]: E0307 01:50:08.483237 3332 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-5b85766d88-58627" Mar 7 01:50:08.485334 kubelet[3332]: E0307 01:50:08.484315 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-5b85766d88-58627_calico-system(e48198d3-aa71-41df-a131-1b93c5c926c5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-5b85766d88-58627_calico-system(e48198d3-aa71-41df-a131-1b93c5c926c5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-5b85766d88-58627" podUID="e48198d3-aa71-41df-a131-1b93c5c926c5" Mar 7 01:50:08.540416 containerd[1817]: time="2026-03-07T01:50:08.540369427Z" level=error msg="Failed to destroy network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.542499 containerd[1817]: time="2026-03-07T01:50:08.542432311Z" level=error msg="encountered an error cleaning up failed sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.542590 containerd[1817]: time="2026-03-07T01:50:08.542517751Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6b776dffb4-p5tkl,Uid:da70d8fb-e941-4544-a89f-11fb7be7f164,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.542855 kubelet[3332]: E0307 01:50:08.542817 3332 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.543072 kubelet[3332]: E0307 01:50:08.543051 3332 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6b776dffb4-p5tkl" Mar 7 01:50:08.543166 kubelet[3332]: E0307 01:50:08.543151 3332 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6b776dffb4-p5tkl" Mar 7 01:50:08.543495 kubelet[3332]: E0307 01:50:08.543269 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-6b776dffb4-p5tkl_calico-system(da70d8fb-e941-4544-a89f-11fb7be7f164)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-6b776dffb4-p5tkl_calico-system(da70d8fb-e941-4544-a89f-11fb7be7f164)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6b776dffb4-p5tkl" podUID="da70d8fb-e941-4544-a89f-11fb7be7f164" Mar 7 01:50:08.564148 containerd[1817]: time="2026-03-07T01:50:08.564103912Z" level=error msg="Failed to destroy network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.564616 containerd[1817]: time="2026-03-07T01:50:08.564584592Z" level=error msg="encountered an error cleaning up failed sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.564733 containerd[1817]: time="2026-03-07T01:50:08.564710553Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-68c448d44-wz8jw,Uid:8de81ec4-1390-4c8d-8e83-a56e0632cd6d,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.565022 kubelet[3332]: E0307 01:50:08.564994 3332 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.565513 kubelet[3332]: E0307 01:50:08.565431 3332 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-68c448d44-wz8jw" Mar 7 01:50:08.565631 kubelet[3332]: E0307 01:50:08.565606 3332 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-68c448d44-wz8jw" Mar 7 01:50:08.565830 kubelet[3332]: E0307 01:50:08.565750 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-68c448d44-wz8jw_calico-system(8de81ec4-1390-4c8d-8e83-a56e0632cd6d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-68c448d44-wz8jw_calico-system(8de81ec4-1390-4c8d-8e83-a56e0632cd6d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-68c448d44-wz8jw" podUID="8de81ec4-1390-4c8d-8e83-a56e0632cd6d" Mar 7 01:50:08.569656 containerd[1817]: time="2026-03-07T01:50:08.569625322Z" level=error msg="Failed to destroy network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.570607 containerd[1817]: time="2026-03-07T01:50:08.570577284Z" level=error msg="encountered an error cleaning up failed sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.570769 containerd[1817]: time="2026-03-07T01:50:08.570729284Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n7zbg,Uid:8cd6b288-aa15-4bc8-b027-e2e3c62e23a5,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.572276 kubelet[3332]: E0307 01:50:08.571014 3332 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.572276 kubelet[3332]: E0307 01:50:08.571060 3332 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-n7zbg" Mar 7 01:50:08.572276 kubelet[3332]: E0307 01:50:08.571076 3332 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-n7zbg" Mar 7 01:50:08.572416 kubelet[3332]: E0307 01:50:08.571111 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-n7zbg_kube-system(8cd6b288-aa15-4bc8-b027-e2e3c62e23a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-n7zbg_kube-system(8cd6b288-aa15-4bc8-b027-e2e3c62e23a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-n7zbg" podUID="8cd6b288-aa15-4bc8-b027-e2e3c62e23a5" Mar 7 01:50:08.583213 containerd[1817]: time="2026-03-07T01:50:08.583105627Z" level=error msg="Failed to destroy network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.583938 containerd[1817]: time="2026-03-07T01:50:08.583909188Z" level=error msg="encountered an error cleaning up failed sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.584077 containerd[1817]: time="2026-03-07T01:50:08.584032589Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hkgbg,Uid:a40182e6-4b51-42f9-8f1a-ad85b1365603,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.584292 kubelet[3332]: E0307 01:50:08.584265 3332 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.584351 kubelet[3332]: E0307 01:50:08.584304 3332 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hkgbg" Mar 7 01:50:08.584351 kubelet[3332]: E0307 01:50:08.584321 3332 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hkgbg" Mar 7 01:50:08.584421 kubelet[3332]: E0307 01:50:08.584365 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-hkgbg_kube-system(a40182e6-4b51-42f9-8f1a-ad85b1365603)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-hkgbg_kube-system(a40182e6-4b51-42f9-8f1a-ad85b1365603)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-hkgbg" podUID="a40182e6-4b51-42f9-8f1a-ad85b1365603" Mar 7 01:50:08.587034 containerd[1817]: time="2026-03-07T01:50:08.586998674Z" level=error msg="Failed to destroy network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.588012 containerd[1817]: time="2026-03-07T01:50:08.587974716Z" level=error msg="encountered an error cleaning up failed sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.588206 containerd[1817]: time="2026-03-07T01:50:08.588174276Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-gb6nc,Uid:f64343b8-a2cd-4274-908a-c1805396f8d4,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.589462 kubelet[3332]: E0307 01:50:08.589284 3332 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.589462 kubelet[3332]: E0307 01:50:08.589343 3332 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57bc98fcf9-gb6nc" Mar 7 01:50:08.589462 kubelet[3332]: E0307 01:50:08.589375 3332 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57bc98fcf9-gb6nc" Mar 7 01:50:08.589603 kubelet[3332]: E0307 01:50:08.589420 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-57bc98fcf9-gb6nc_calico-system(f64343b8-a2cd-4274-908a-c1805396f8d4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-57bc98fcf9-gb6nc_calico-system(f64343b8-a2cd-4274-908a-c1805396f8d4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-57bc98fcf9-gb6nc" podUID="f64343b8-a2cd-4274-908a-c1805396f8d4" Mar 7 01:50:08.590571 containerd[1817]: time="2026-03-07T01:50:08.589738479Z" level=error msg="Failed to destroy network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.591112 containerd[1817]: time="2026-03-07T01:50:08.590931602Z" level=error msg="encountered an error cleaning up failed sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.591112 containerd[1817]: time="2026-03-07T01:50:08.590989522Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-qfvhc,Uid:58c58ebb-a76c-40e3-bec1-27853dcf0611,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.591268 kubelet[3332]: E0307 01:50:08.591174 3332 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.591268 kubelet[3332]: E0307 01:50:08.591218 3332 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57bc98fcf9-qfvhc" Mar 7 01:50:08.591268 kubelet[3332]: E0307 01:50:08.591234 3332 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-57bc98fcf9-qfvhc" Mar 7 01:50:08.591386 kubelet[3332]: E0307 01:50:08.591274 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-57bc98fcf9-qfvhc_calico-system(58c58ebb-a76c-40e3-bec1-27853dcf0611)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-57bc98fcf9-qfvhc_calico-system(58c58ebb-a76c-40e3-bec1-27853dcf0611)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-57bc98fcf9-qfvhc" podUID="58c58ebb-a76c-40e3-bec1-27853dcf0611" Mar 7 01:50:08.679853 kubelet[3332]: I0307 01:50:08.679502 3332 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:08.680600 containerd[1817]: time="2026-03-07T01:50:08.680274328Z" level=info msg="StopPodSandbox for \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\"" Mar 7 01:50:08.685923 kubelet[3332]: I0307 01:50:08.685898 3332 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:08.686597 containerd[1817]: time="2026-03-07T01:50:08.686412299Z" level=info msg="StopPodSandbox for \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\"" Mar 7 01:50:08.687034 containerd[1817]: time="2026-03-07T01:50:08.686822620Z" level=info msg="Ensure that sandbox ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8 in task-service has been cleanup successfully" Mar 7 01:50:08.690286 kubelet[3332]: I0307 01:50:08.689164 3332 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:08.691562 containerd[1817]: time="2026-03-07T01:50:08.690138466Z" level=info msg="StopPodSandbox for \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\"" Mar 7 01:50:08.692933 containerd[1817]: time="2026-03-07T01:50:08.692905911Z" level=info msg="Ensure that sandbox 143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5 in task-service has been cleanup successfully" Mar 7 01:50:08.696224 containerd[1817]: time="2026-03-07T01:50:08.696189318Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 7 01:50:08.701058 kubelet[3332]: I0307 01:50:08.700605 3332 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:08.701538 containerd[1817]: time="2026-03-07T01:50:08.701514488Z" level=info msg="Ensure that sandbox 6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6 in task-service has been cleanup successfully" Mar 7 01:50:08.703770 containerd[1817]: time="2026-03-07T01:50:08.703519251Z" level=info msg="StopPodSandbox for \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\"" Mar 7 01:50:08.704120 containerd[1817]: time="2026-03-07T01:50:08.704084172Z" level=info msg="Ensure that sandbox a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e in task-service has been cleanup successfully" Mar 7 01:50:08.710033 kubelet[3332]: I0307 01:50:08.709512 3332 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:08.710935 containerd[1817]: time="2026-03-07T01:50:08.710214704Z" level=info msg="StopPodSandbox for \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\"" Mar 7 01:50:08.710935 containerd[1817]: time="2026-03-07T01:50:08.710351984Z" level=info msg="Ensure that sandbox f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab in task-service has been cleanup successfully" Mar 7 01:50:08.712716 kubelet[3332]: I0307 01:50:08.712693 3332 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:08.713887 containerd[1817]: time="2026-03-07T01:50:08.713741870Z" level=info msg="StopPodSandbox for \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\"" Mar 7 01:50:08.715338 containerd[1817]: time="2026-03-07T01:50:08.715315793Z" level=info msg="Ensure that sandbox b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853 in task-service has been cleanup successfully" Mar 7 01:50:08.721580 kubelet[3332]: I0307 01:50:08.721564 3332 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:08.724618 containerd[1817]: time="2026-03-07T01:50:08.724594771Z" level=info msg="StopPodSandbox for \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\"" Mar 7 01:50:08.725130 containerd[1817]: time="2026-03-07T01:50:08.725107491Z" level=info msg="Ensure that sandbox 619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2 in task-service has been cleanup successfully" Mar 7 01:50:08.757993 containerd[1817]: time="2026-03-07T01:50:08.757760512Z" level=info msg="CreateContainer within sandbox \"c8948280b4f28d6c4f4ab9ad499bad0f424439d53be299f0064f311fbee90d96\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"e5f2066423321030319835910929c0b59b60fb36d6c178bdbc6d9b7b4cd728fc\"" Mar 7 01:50:08.759573 containerd[1817]: time="2026-03-07T01:50:08.759544276Z" level=info msg="StartContainer for \"e5f2066423321030319835910929c0b59b60fb36d6c178bdbc6d9b7b4cd728fc\"" Mar 7 01:50:08.844690 containerd[1817]: time="2026-03-07T01:50:08.844623154Z" level=error msg="StopPodSandbox for \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\" failed" error="failed to destroy network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.845562 kubelet[3332]: E0307 01:50:08.845220 3332 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:08.845562 kubelet[3332]: E0307 01:50:08.845283 3332 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e"} Mar 7 01:50:08.845562 kubelet[3332]: E0307 01:50:08.845335 3332 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f64343b8-a2cd-4274-908a-c1805396f8d4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:50:08.845562 kubelet[3332]: E0307 01:50:08.845356 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f64343b8-a2cd-4274-908a-c1805396f8d4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-57bc98fcf9-gb6nc" podUID="f64343b8-a2cd-4274-908a-c1805396f8d4" Mar 7 01:50:08.856776 containerd[1817]: time="2026-03-07T01:50:08.856625176Z" level=error msg="StopPodSandbox for \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\" failed" error="failed to destroy network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.857059 kubelet[3332]: E0307 01:50:08.856874 3332 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:08.857291 kubelet[3332]: E0307 01:50:08.857142 3332 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8"} Mar 7 01:50:08.857291 kubelet[3332]: E0307 01:50:08.857177 3332 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"da70d8fb-e941-4544-a89f-11fb7be7f164\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:50:08.857291 kubelet[3332]: E0307 01:50:08.857251 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"da70d8fb-e941-4544-a89f-11fb7be7f164\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6b776dffb4-p5tkl" podUID="da70d8fb-e941-4544-a89f-11fb7be7f164" Mar 7 01:50:08.873308 containerd[1817]: time="2026-03-07T01:50:08.872216806Z" level=error msg="StopPodSandbox for \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\" failed" error="failed to destroy network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.873308 containerd[1817]: time="2026-03-07T01:50:08.873164927Z" level=error msg="StopPodSandbox for \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\" failed" error="failed to destroy network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.873420 kubelet[3332]: E0307 01:50:08.872940 3332 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:08.873420 kubelet[3332]: E0307 01:50:08.872980 3332 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab"} Mar 7 01:50:08.873420 kubelet[3332]: E0307 01:50:08.873009 3332 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8de81ec4-1390-4c8d-8e83-a56e0632cd6d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:50:08.873420 kubelet[3332]: E0307 01:50:08.873031 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8de81ec4-1390-4c8d-8e83-a56e0632cd6d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-68c448d44-wz8jw" podUID="8de81ec4-1390-4c8d-8e83-a56e0632cd6d" Mar 7 01:50:08.873828 kubelet[3332]: E0307 01:50:08.873284 3332 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:08.873828 kubelet[3332]: E0307 01:50:08.873624 3332 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6"} Mar 7 01:50:08.873828 kubelet[3332]: E0307 01:50:08.873650 3332 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"58c58ebb-a76c-40e3-bec1-27853dcf0611\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:50:08.873828 kubelet[3332]: E0307 01:50:08.873680 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"58c58ebb-a76c-40e3-bec1-27853dcf0611\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-57bc98fcf9-qfvhc" podUID="58c58ebb-a76c-40e3-bec1-27853dcf0611" Mar 7 01:50:08.877547 containerd[1817]: time="2026-03-07T01:50:08.876596814Z" level=error msg="StopPodSandbox for \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\" failed" error="failed to destroy network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.877625 kubelet[3332]: E0307 01:50:08.876746 3332 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:08.877625 kubelet[3332]: E0307 01:50:08.876788 3332 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2"} Mar 7 01:50:08.877625 kubelet[3332]: E0307 01:50:08.876812 3332 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"e48198d3-aa71-41df-a131-1b93c5c926c5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:50:08.877625 kubelet[3332]: E0307 01:50:08.876831 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"e48198d3-aa71-41df-a131-1b93c5c926c5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-5b85766d88-58627" podUID="e48198d3-aa71-41df-a131-1b93c5c926c5" Mar 7 01:50:08.879324 containerd[1817]: time="2026-03-07T01:50:08.878995698Z" level=error msg="StopPodSandbox for \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\" failed" error="failed to destroy network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.879418 kubelet[3332]: E0307 01:50:08.879209 3332 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:08.879418 kubelet[3332]: E0307 01:50:08.879251 3332 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5"} Mar 7 01:50:08.879418 kubelet[3332]: E0307 01:50:08.879273 3332 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"a40182e6-4b51-42f9-8f1a-ad85b1365603\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:50:08.879418 kubelet[3332]: E0307 01:50:08.879292 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"a40182e6-4b51-42f9-8f1a-ad85b1365603\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-hkgbg" podUID="a40182e6-4b51-42f9-8f1a-ad85b1365603" Mar 7 01:50:08.883021 containerd[1817]: time="2026-03-07T01:50:08.882976346Z" level=error msg="StopPodSandbox for \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\" failed" error="failed to destroy network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 7 01:50:08.884994 kubelet[3332]: E0307 01:50:08.883116 3332 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:08.884994 kubelet[3332]: E0307 01:50:08.883149 3332 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853"} Mar 7 01:50:08.884994 kubelet[3332]: E0307 01:50:08.883171 3332 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 7 01:50:08.884994 kubelet[3332]: E0307 01:50:08.883189 3332 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-n7zbg" podUID="8cd6b288-aa15-4bc8-b027-e2e3c62e23a5" Mar 7 01:50:08.887989 containerd[1817]: time="2026-03-07T01:50:08.887958955Z" level=info msg="StartContainer for \"e5f2066423321030319835910929c0b59b60fb36d6c178bdbc6d9b7b4cd728fc\" returns successfully" Mar 7 01:50:09.550899 containerd[1817]: time="2026-03-07T01:50:09.550858710Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-cz2s4,Uid:c189a211-99d4-4133-a1d8-fbd792f2c0a5,Namespace:calico-system,Attempt:0,}" Mar 7 01:50:09.694834 systemd-networkd[1378]: cali6d797cfdd5c: Link UP Mar 7 01:50:09.695838 systemd-networkd[1378]: cali6d797cfdd5c: Gained carrier Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.608 [ERROR][4523] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.625 [INFO][4523] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0 csi-node-driver- calico-system c189a211-99d4-4133-a1d8-fbd792f2c0a5 720 0 2026-03-07 01:49:47 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6d9d697c7c k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af csi-node-driver-cz2s4 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali6d797cfdd5c [] [] }} ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.626 [INFO][4523] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.648 [INFO][4535] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" HandleID="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.656 [INFO][4535] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" HandleID="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fb7a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"csi-node-driver-cz2s4", "timestamp":"2026-03-07 01:50:09.648041131 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40003071e0)} Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.656 [INFO][4535] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.656 [INFO][4535] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.656 [INFO][4535] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.658 [INFO][4535] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.661 [INFO][4535] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.668 [INFO][4535] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.670 [INFO][4535] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.672 [INFO][4535] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.672 [INFO][4535] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.673 [INFO][4535] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082 Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.677 [INFO][4535] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.685 [INFO][4535] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.65/26] block=192.168.112.64/26 handle="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.685 [INFO][4535] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.65/26] handle="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.685 [INFO][4535] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:09.720009 containerd[1817]: 2026-03-07 01:50:09.685 [INFO][4535] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.65/26] IPv6=[] ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" HandleID="k8s-pod-network.0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" Mar 7 01:50:09.720554 containerd[1817]: 2026-03-07 01:50:09.689 [INFO][4523] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"c189a211-99d4-4133-a1d8-fbd792f2c0a5", ResourceVersion:"720", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"csi-node-driver-cz2s4", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.112.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6d797cfdd5c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:09.720554 containerd[1817]: 2026-03-07 01:50:09.689 [INFO][4523] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.65/32] ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" Mar 7 01:50:09.720554 containerd[1817]: 2026-03-07 01:50:09.689 [INFO][4523] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6d797cfdd5c ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" Mar 7 01:50:09.720554 containerd[1817]: 2026-03-07 01:50:09.696 [INFO][4523] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" Mar 7 01:50:09.720554 containerd[1817]: 2026-03-07 01:50:09.697 [INFO][4523] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"c189a211-99d4-4133-a1d8-fbd792f2c0a5", ResourceVersion:"720", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6d9d697c7c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082", Pod:"csi-node-driver-cz2s4", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.112.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6d797cfdd5c", MAC:"12:98:62:96:0f:78", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:09.720554 containerd[1817]: 2026-03-07 01:50:09.716 [INFO][4523] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082" Namespace="calico-system" Pod="csi-node-driver-cz2s4" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-csi--node--driver--cz2s4-eth0" Mar 7 01:50:09.732889 containerd[1817]: time="2026-03-07T01:50:09.730532125Z" level=info msg="StopPodSandbox for \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\"" Mar 7 01:50:09.739014 containerd[1817]: time="2026-03-07T01:50:09.738921820Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:09.739014 containerd[1817]: time="2026-03-07T01:50:09.738975860Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:09.739014 containerd[1817]: time="2026-03-07T01:50:09.738990620Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:09.739174 containerd[1817]: time="2026-03-07T01:50:09.739065340Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:09.792808 containerd[1817]: time="2026-03-07T01:50:09.792760240Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-cz2s4,Uid:c189a211-99d4-4133-a1d8-fbd792f2c0a5,Namespace:calico-system,Attempt:0,} returns sandbox id \"0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082\"" Mar 7 01:50:09.794571 containerd[1817]: time="2026-03-07T01:50:09.794547084Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\"" Mar 7 01:50:09.824205 kubelet[3332]: I0307 01:50:09.821490 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-5whnz" podStartSLOduration=5.781265713 podStartE2EDuration="23.821472734s" podCreationTimestamp="2026-03-07 01:49:46 +0000 UTC" firstStartedPulling="2026-03-07 01:49:47.40472556 +0000 UTC m=+18.952327596" lastFinishedPulling="2026-03-07 01:50:05.444932581 +0000 UTC m=+36.992534617" observedRunningTime="2026-03-07 01:50:09.753012086 +0000 UTC m=+41.300614122" watchObservedRunningTime="2026-03-07 01:50:09.821472734 +0000 UTC m=+41.369074770" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.825 [INFO][4585] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.825 [INFO][4585] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" iface="eth0" netns="/var/run/netns/cni-c3b47696-e2a7-6ec5-ea8e-68aecad8eed3" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.826 [INFO][4585] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" iface="eth0" netns="/var/run/netns/cni-c3b47696-e2a7-6ec5-ea8e-68aecad8eed3" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.826 [INFO][4585] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" iface="eth0" netns="/var/run/netns/cni-c3b47696-e2a7-6ec5-ea8e-68aecad8eed3" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.826 [INFO][4585] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.826 [INFO][4585] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.844 [INFO][4610] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.844 [INFO][4610] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.844 [INFO][4610] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.852 [WARNING][4610] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.852 [INFO][4610] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.853 [INFO][4610] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:09.857409 containerd[1817]: 2026-03-07 01:50:09.855 [INFO][4585] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:09.860489 containerd[1817]: time="2026-03-07T01:50:09.858199482Z" level=info msg="TearDown network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\" successfully" Mar 7 01:50:09.860489 containerd[1817]: time="2026-03-07T01:50:09.858232282Z" level=info msg="StopPodSandbox for \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\" returns successfully" Mar 7 01:50:09.860492 systemd[1]: run-netns-cni\x2dc3b47696\x2de2a7\x2d6ec5\x2dea8e\x2d68aecad8eed3.mount: Deactivated successfully. Mar 7 01:50:09.910730 kubelet[3332]: I0307 01:50:09.910690 3332 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkt9q\" (UniqueName: \"kubernetes.io/projected/da70d8fb-e941-4544-a89f-11fb7be7f164-kube-api-access-lkt9q\") pod \"da70d8fb-e941-4544-a89f-11fb7be7f164\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " Mar 7 01:50:09.910846 kubelet[3332]: I0307 01:50:09.910740 3332 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-backend-key-pair\") pod \"da70d8fb-e941-4544-a89f-11fb7be7f164\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " Mar 7 01:50:09.910846 kubelet[3332]: I0307 01:50:09.910778 3332 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-nginx-config\") pod \"da70d8fb-e941-4544-a89f-11fb7be7f164\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " Mar 7 01:50:09.910846 kubelet[3332]: I0307 01:50:09.910796 3332 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-ca-bundle\") pod \"da70d8fb-e941-4544-a89f-11fb7be7f164\" (UID: \"da70d8fb-e941-4544-a89f-11fb7be7f164\") " Mar 7 01:50:09.912868 kubelet[3332]: I0307 01:50:09.911537 3332 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "da70d8fb-e941-4544-a89f-11fb7be7f164" (UID: "da70d8fb-e941-4544-a89f-11fb7be7f164"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 7 01:50:09.912868 kubelet[3332]: I0307 01:50:09.911782 3332 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-nginx-config" (OuterVolumeSpecName: "nginx-config") pod "da70d8fb-e941-4544-a89f-11fb7be7f164" (UID: "da70d8fb-e941-4544-a89f-11fb7be7f164"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 7 01:50:09.916179 kubelet[3332]: I0307 01:50:09.916152 3332 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da70d8fb-e941-4544-a89f-11fb7be7f164-kube-api-access-lkt9q" (OuterVolumeSpecName: "kube-api-access-lkt9q") pod "da70d8fb-e941-4544-a89f-11fb7be7f164" (UID: "da70d8fb-e941-4544-a89f-11fb7be7f164"). InnerVolumeSpecName "kube-api-access-lkt9q". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 7 01:50:09.917457 systemd[1]: var-lib-kubelet-pods-da70d8fb\x2de941\x2d4544\x2da89f\x2d11fb7be7f164-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dlkt9q.mount: Deactivated successfully. Mar 7 01:50:09.917657 systemd[1]: var-lib-kubelet-pods-da70d8fb\x2de941\x2d4544\x2da89f\x2d11fb7be7f164-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Mar 7 01:50:09.917765 kubelet[3332]: I0307 01:50:09.916201 3332 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "da70d8fb-e941-4544-a89f-11fb7be7f164" (UID: "da70d8fb-e941-4544-a89f-11fb7be7f164"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 7 01:50:10.011827 kubelet[3332]: I0307 01:50:10.011791 3332 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lkt9q\" (UniqueName: \"kubernetes.io/projected/da70d8fb-e941-4544-a89f-11fb7be7f164-kube-api-access-lkt9q\") on node \"ci-4081.3.6-n-e8b1fc37af\" DevicePath \"\"" Mar 7 01:50:10.012000 kubelet[3332]: I0307 01:50:10.011988 3332 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-backend-key-pair\") on node \"ci-4081.3.6-n-e8b1fc37af\" DevicePath \"\"" Mar 7 01:50:10.012061 kubelet[3332]: I0307 01:50:10.012052 3332 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-nginx-config\") on node \"ci-4081.3.6-n-e8b1fc37af\" DevicePath \"\"" Mar 7 01:50:10.012119 kubelet[3332]: I0307 01:50:10.012109 3332 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da70d8fb-e941-4544-a89f-11fb7be7f164-whisker-ca-bundle\") on node \"ci-4081.3.6-n-e8b1fc37af\" DevicePath \"\"" Mar 7 01:50:10.508464 kernel: calico-node[4636]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Mar 7 01:50:10.920723 kubelet[3332]: I0307 01:50:10.920677 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/b341ab2c-6064-44bd-857d-12b9fa393a56-nginx-config\") pod \"whisker-56db756f69-ksrx8\" (UID: \"b341ab2c-6064-44bd-857d-12b9fa393a56\") " pod="calico-system/whisker-56db756f69-ksrx8" Mar 7 01:50:10.921130 kubelet[3332]: I0307 01:50:10.920733 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/b341ab2c-6064-44bd-857d-12b9fa393a56-whisker-backend-key-pair\") pod \"whisker-56db756f69-ksrx8\" (UID: \"b341ab2c-6064-44bd-857d-12b9fa393a56\") " pod="calico-system/whisker-56db756f69-ksrx8" Mar 7 01:50:10.921130 kubelet[3332]: I0307 01:50:10.920756 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcsrg\" (UniqueName: \"kubernetes.io/projected/b341ab2c-6064-44bd-857d-12b9fa393a56-kube-api-access-pcsrg\") pod \"whisker-56db756f69-ksrx8\" (UID: \"b341ab2c-6064-44bd-857d-12b9fa393a56\") " pod="calico-system/whisker-56db756f69-ksrx8" Mar 7 01:50:10.921130 kubelet[3332]: I0307 01:50:10.920776 3332 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b341ab2c-6064-44bd-857d-12b9fa393a56-whisker-ca-bundle\") pod \"whisker-56db756f69-ksrx8\" (UID: \"b341ab2c-6064-44bd-857d-12b9fa393a56\") " pod="calico-system/whisker-56db756f69-ksrx8" Mar 7 01:50:11.058992 systemd-networkd[1378]: vxlan.calico: Link UP Mar 7 01:50:11.058999 systemd-networkd[1378]: vxlan.calico: Gained carrier Mar 7 01:50:11.129217 containerd[1817]: time="2026-03-07T01:50:11.128884290Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-56db756f69-ksrx8,Uid:b341ab2c-6064-44bd-857d-12b9fa393a56,Namespace:calico-system,Attempt:0,}" Mar 7 01:50:11.137022 systemd-networkd[1378]: cali6d797cfdd5c: Gained IPv6LL Mar 7 01:50:11.346455 systemd-networkd[1378]: cali8badd65fca2: Link UP Mar 7 01:50:11.350655 systemd-networkd[1378]: cali8badd65fca2: Gained carrier Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.252 [INFO][4795] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0 whisker-56db756f69- calico-system b341ab2c-6064-44bd-857d-12b9fa393a56 943 0 2026-03-07 01:50:10 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:56db756f69 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af whisker-56db756f69-ksrx8 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali8badd65fca2 [] [] }} ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.253 [INFO][4795] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.280 [INFO][4813] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" HandleID="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.291 [INFO][4813] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" HandleID="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000273270), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"whisker-56db756f69-ksrx8", "timestamp":"2026-03-07 01:50:11.280024971 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000326f20)} Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.291 [INFO][4813] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.291 [INFO][4813] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.291 [INFO][4813] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.293 [INFO][4813] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.297 [INFO][4813] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.301 [INFO][4813] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.303 [INFO][4813] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.306 [INFO][4813] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.306 [INFO][4813] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.312 [INFO][4813] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.318 [INFO][4813] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.329 [INFO][4813] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.66/26] block=192.168.112.64/26 handle="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.329 [INFO][4813] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.66/26] handle="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.330 [INFO][4813] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:11.392815 containerd[1817]: 2026-03-07 01:50:11.330 [INFO][4813] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.66/26] IPv6=[] ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" HandleID="k8s-pod-network.e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" Mar 7 01:50:11.393374 containerd[1817]: 2026-03-07 01:50:11.336 [INFO][4795] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0", GenerateName:"whisker-56db756f69-", Namespace:"calico-system", SelfLink:"", UID:"b341ab2c-6064-44bd-857d-12b9fa393a56", ResourceVersion:"943", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 50, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"56db756f69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"whisker-56db756f69-ksrx8", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.112.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali8badd65fca2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:11.393374 containerd[1817]: 2026-03-07 01:50:11.337 [INFO][4795] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.66/32] ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" Mar 7 01:50:11.393374 containerd[1817]: 2026-03-07 01:50:11.338 [INFO][4795] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8badd65fca2 ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" Mar 7 01:50:11.393374 containerd[1817]: 2026-03-07 01:50:11.362 [INFO][4795] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" Mar 7 01:50:11.393374 containerd[1817]: 2026-03-07 01:50:11.363 [INFO][4795] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0", GenerateName:"whisker-56db756f69-", Namespace:"calico-system", SelfLink:"", UID:"b341ab2c-6064-44bd-857d-12b9fa393a56", ResourceVersion:"943", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 50, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"56db756f69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e", Pod:"whisker-56db756f69-ksrx8", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.112.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali8badd65fca2", MAC:"fe:c0:fc:76:76:2a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:11.393374 containerd[1817]: 2026-03-07 01:50:11.379 [INFO][4795] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e" Namespace="calico-system" Pod="whisker-56db756f69-ksrx8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--56db756f69--ksrx8-eth0" Mar 7 01:50:11.467298 containerd[1817]: time="2026-03-07T01:50:11.461393429Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:11.467298 containerd[1817]: time="2026-03-07T01:50:11.461468429Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:11.467298 containerd[1817]: time="2026-03-07T01:50:11.461485029Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:11.467298 containerd[1817]: time="2026-03-07T01:50:11.461571509Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:11.519953 containerd[1817]: time="2026-03-07T01:50:11.517699654Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:11.522269 containerd[1817]: time="2026-03-07T01:50:11.522220782Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.4: active requests=0, bytes read=8261497" Mar 7 01:50:11.526752 containerd[1817]: time="2026-03-07T01:50:11.526710591Z" level=info msg="ImageCreate event name:\"sha256:9cb4086a1b408b52c6b14e0b81520060e1766ee0243508d29d8a53c7b518051f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:11.528027 containerd[1817]: time="2026-03-07T01:50:11.527994993Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-56db756f69-ksrx8,Uid:b341ab2c-6064-44bd-857d-12b9fa393a56,Namespace:calico-system,Attempt:0,} returns sandbox id \"e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e\"" Mar 7 01:50:11.540326 containerd[1817]: time="2026-03-07T01:50:11.540290736Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:11.541502 containerd[1817]: time="2026-03-07T01:50:11.541476018Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.4\" with image id \"sha256:9cb4086a1b408b52c6b14e0b81520060e1766ee0243508d29d8a53c7b518051f\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\", size \"9659022\" in 1.746793094s" Mar 7 01:50:11.541605 containerd[1817]: time="2026-03-07T01:50:11.541590139Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\" returns image reference \"sha256:9cb4086a1b408b52c6b14e0b81520060e1766ee0243508d29d8a53c7b518051f\"" Mar 7 01:50:11.544724 containerd[1817]: time="2026-03-07T01:50:11.544695104Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\"" Mar 7 01:50:11.551636 containerd[1817]: time="2026-03-07T01:50:11.551600997Z" level=info msg="CreateContainer within sandbox \"0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 7 01:50:11.599981 containerd[1817]: time="2026-03-07T01:50:11.598923085Z" level=info msg="CreateContainer within sandbox \"0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"c25328bbd0d91ca41078bfdd68a3d0f2db2e0ba2b2ca095fdd9c2887f2e49c2b\"" Mar 7 01:50:11.600633 containerd[1817]: time="2026-03-07T01:50:11.600604688Z" level=info msg="StartContainer for \"c25328bbd0d91ca41078bfdd68a3d0f2db2e0ba2b2ca095fdd9c2887f2e49c2b\"" Mar 7 01:50:11.655184 containerd[1817]: time="2026-03-07T01:50:11.655146750Z" level=info msg="StartContainer for \"c25328bbd0d91ca41078bfdd68a3d0f2db2e0ba2b2ca095fdd9c2887f2e49c2b\" returns successfully" Mar 7 01:50:12.416570 systemd-networkd[1378]: cali8badd65fca2: Gained IPv6LL Mar 7 01:50:12.544549 systemd-networkd[1378]: vxlan.calico: Gained IPv6LL Mar 7 01:50:12.551241 kubelet[3332]: I0307 01:50:12.551183 3332 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da70d8fb-e941-4544-a89f-11fb7be7f164" path="/var/lib/kubelet/pods/da70d8fb-e941-4544-a89f-11fb7be7f164/volumes" Mar 7 01:50:13.052225 containerd[1817]: time="2026-03-07T01:50:13.052172471Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:13.055698 containerd[1817]: time="2026-03-07T01:50:13.055541157Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.4: active requests=0, bytes read=5882804" Mar 7 01:50:13.058909 containerd[1817]: time="2026-03-07T01:50:13.058624923Z" level=info msg="ImageCreate event name:\"sha256:51af4e9dcdb93e51b26a4a6f99272ec2df8de1aef256bb746f2c7c844b8e7b2c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:13.064159 containerd[1817]: time="2026-03-07T01:50:13.064098693Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:13.065047 containerd[1817]: time="2026-03-07T01:50:13.064953535Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.4\" with image id \"sha256:51af4e9dcdb93e51b26a4a6f99272ec2df8de1aef256bb746f2c7c844b8e7b2c\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\", size \"7280321\" in 1.520226471s" Mar 7 01:50:13.065047 containerd[1817]: time="2026-03-07T01:50:13.064995855Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\" returns image reference \"sha256:51af4e9dcdb93e51b26a4a6f99272ec2df8de1aef256bb746f2c7c844b8e7b2c\"" Mar 7 01:50:13.066575 containerd[1817]: time="2026-03-07T01:50:13.066549298Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\"" Mar 7 01:50:13.072849 containerd[1817]: time="2026-03-07T01:50:13.072670869Z" level=info msg="CreateContainer within sandbox \"e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Mar 7 01:50:13.132917 containerd[1817]: time="2026-03-07T01:50:13.132875221Z" level=info msg="CreateContainer within sandbox \"e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"624037ad217c609f5e94d26c9eb0610b611618fdd7264b8726dffd631afd04c8\"" Mar 7 01:50:13.133619 containerd[1817]: time="2026-03-07T01:50:13.133586862Z" level=info msg="StartContainer for \"624037ad217c609f5e94d26c9eb0610b611618fdd7264b8726dffd631afd04c8\"" Mar 7 01:50:13.161166 systemd[1]: run-containerd-runc-k8s.io-624037ad217c609f5e94d26c9eb0610b611618fdd7264b8726dffd631afd04c8-runc.dcWgFM.mount: Deactivated successfully. Mar 7 01:50:13.196123 containerd[1817]: time="2026-03-07T01:50:13.195779578Z" level=info msg="StartContainer for \"624037ad217c609f5e94d26c9eb0610b611618fdd7264b8726dffd631afd04c8\" returns successfully" Mar 7 01:50:14.600509 containerd[1817]: time="2026-03-07T01:50:14.600434827Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:14.604355 containerd[1817]: time="2026-03-07T01:50:14.604322955Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4: active requests=0, bytes read=13766291" Mar 7 01:50:14.607702 containerd[1817]: time="2026-03-07T01:50:14.607659881Z" level=info msg="ImageCreate event name:\"sha256:8195c49a3b504e7ef58a8fc9a0e9ae66ae6ae90ef4998c04591be9588e8fa07e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:14.613033 containerd[1817]: time="2026-03-07T01:50:14.612949371Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:14.614019 containerd[1817]: time="2026-03-07T01:50:14.613874612Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" with image id \"sha256:8195c49a3b504e7ef58a8fc9a0e9ae66ae6ae90ef4998c04591be9588e8fa07e\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\", size \"15163768\" in 1.547293754s" Mar 7 01:50:14.614019 containerd[1817]: time="2026-03-07T01:50:14.613906572Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" returns image reference \"sha256:8195c49a3b504e7ef58a8fc9a0e9ae66ae6ae90ef4998c04591be9588e8fa07e\"" Mar 7 01:50:14.615494 containerd[1817]: time="2026-03-07T01:50:14.615467655Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\"" Mar 7 01:50:14.622314 containerd[1817]: time="2026-03-07T01:50:14.622195868Z" level=info msg="CreateContainer within sandbox \"0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 7 01:50:14.662669 containerd[1817]: time="2026-03-07T01:50:14.662625423Z" level=info msg="CreateContainer within sandbox \"0bec004667afb7a006bd84ee0aa8a2dab7748d7144ab030695f8a10c1a161082\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"5ad8b2965d099a24e2817f17ad45358c14d3d0580804b37a59de048847ba6e98\"" Mar 7 01:50:14.663508 containerd[1817]: time="2026-03-07T01:50:14.663465024Z" level=info msg="StartContainer for \"5ad8b2965d099a24e2817f17ad45358c14d3d0580804b37a59de048847ba6e98\"" Mar 7 01:50:14.715265 containerd[1817]: time="2026-03-07T01:50:14.715227601Z" level=info msg="StartContainer for \"5ad8b2965d099a24e2817f17ad45358c14d3d0580804b37a59de048847ba6e98\" returns successfully" Mar 7 01:50:15.558079 kubelet[3332]: I0307 01:50:15.557613 3332 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:50:15.664793 kubelet[3332]: I0307 01:50:15.664347 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-cz2s4" podStartSLOduration=23.843234712 podStartE2EDuration="28.664319004s" podCreationTimestamp="2026-03-07 01:49:47 +0000 UTC" firstStartedPulling="2026-03-07 01:50:09.794054563 +0000 UTC m=+41.341656599" lastFinishedPulling="2026-03-07 01:50:14.615138855 +0000 UTC m=+46.162740891" observedRunningTime="2026-03-07 01:50:14.76320049 +0000 UTC m=+46.310802526" watchObservedRunningTime="2026-03-07 01:50:15.664319004 +0000 UTC m=+47.211921040" Mar 7 01:50:15.668156 kubelet[3332]: I0307 01:50:15.666971 3332 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 7 01:50:15.668156 kubelet[3332]: I0307 01:50:15.667102 3332 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 7 01:50:16.368551 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount953607008.mount: Deactivated successfully. Mar 7 01:50:16.425583 containerd[1817]: time="2026-03-07T01:50:16.425538258Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:16.429633 containerd[1817]: time="2026-03-07T01:50:16.429447505Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.4: active requests=0, bytes read=16426594" Mar 7 01:50:16.434312 containerd[1817]: time="2026-03-07T01:50:16.434238194Z" level=info msg="ImageCreate event name:\"sha256:19fab8e13a4d97732973f299576e43f89b889ceff6e3768f711f30e6ace1c662\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:16.439675 containerd[1817]: time="2026-03-07T01:50:16.439629564Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:16.440526 containerd[1817]: time="2026-03-07T01:50:16.440375565Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" with image id \"sha256:19fab8e13a4d97732973f299576e43f89b889ceff6e3768f711f30e6ace1c662\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\", size \"16426424\" in 1.82487743s" Mar 7 01:50:16.440526 containerd[1817]: time="2026-03-07T01:50:16.440406126Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" returns image reference \"sha256:19fab8e13a4d97732973f299576e43f89b889ceff6e3768f711f30e6ace1c662\"" Mar 7 01:50:16.447552 containerd[1817]: time="2026-03-07T01:50:16.447521099Z" level=info msg="CreateContainer within sandbox \"e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Mar 7 01:50:16.484011 containerd[1817]: time="2026-03-07T01:50:16.483948366Z" level=info msg="CreateContainer within sandbox \"e3851678a9aa51a2825fa4a42799269a4c7d73a2db98379293d4ca78d3b6023e\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"7c454a74e7222e69ebf941c8a39db4ecbf9a5f67006db9692d6d5747a7b49769\"" Mar 7 01:50:16.489700 containerd[1817]: time="2026-03-07T01:50:16.489519737Z" level=info msg="StartContainer for \"7c454a74e7222e69ebf941c8a39db4ecbf9a5f67006db9692d6d5747a7b49769\"" Mar 7 01:50:16.559177 containerd[1817]: time="2026-03-07T01:50:16.559084586Z" level=info msg="StartContainer for \"7c454a74e7222e69ebf941c8a39db4ecbf9a5f67006db9692d6d5747a7b49769\" returns successfully" Mar 7 01:50:16.770178 kubelet[3332]: I0307 01:50:16.770040 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-56db756f69-ksrx8" podStartSLOduration=1.858107647 podStartE2EDuration="6.770025938s" podCreationTimestamp="2026-03-07 01:50:10 +0000 UTC" firstStartedPulling="2026-03-07 01:50:11.529507556 +0000 UTC m=+43.077109592" lastFinishedPulling="2026-03-07 01:50:16.441425847 +0000 UTC m=+47.989027883" observedRunningTime="2026-03-07 01:50:16.768869656 +0000 UTC m=+48.316471692" watchObservedRunningTime="2026-03-07 01:50:16.770025938 +0000 UTC m=+48.317627974" Mar 7 01:50:19.548895 containerd[1817]: time="2026-03-07T01:50:19.548853060Z" level=info msg="StopPodSandbox for \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\"" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.599 [INFO][5136] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.599 [INFO][5136] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" iface="eth0" netns="/var/run/netns/cni-ba4a16d3-a2c7-632c-2851-fb969ebb2f1d" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.600 [INFO][5136] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" iface="eth0" netns="/var/run/netns/cni-ba4a16d3-a2c7-632c-2851-fb969ebb2f1d" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.601 [INFO][5136] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" iface="eth0" netns="/var/run/netns/cni-ba4a16d3-a2c7-632c-2851-fb969ebb2f1d" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.601 [INFO][5136] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.601 [INFO][5136] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.618 [INFO][5144] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.618 [INFO][5144] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.618 [INFO][5144] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.626 [WARNING][5144] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.626 [INFO][5144] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.627 [INFO][5144] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:19.631359 containerd[1817]: 2026-03-07 01:50:19.629 [INFO][5136] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:19.633631 containerd[1817]: time="2026-03-07T01:50:19.633588098Z" level=info msg="TearDown network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\" successfully" Mar 7 01:50:19.633631 containerd[1817]: time="2026-03-07T01:50:19.633629938Z" level=info msg="StopPodSandbox for \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\" returns successfully" Mar 7 01:50:19.634655 containerd[1817]: time="2026-03-07T01:50:19.634625980Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-58627,Uid:e48198d3-aa71-41df-a131-1b93c5c926c5,Namespace:calico-system,Attempt:1,}" Mar 7 01:50:19.635373 systemd[1]: run-netns-cni\x2dba4a16d3\x2da2c7\x2d632c\x2d2851\x2dfb969ebb2f1d.mount: Deactivated successfully. Mar 7 01:50:19.776791 systemd-networkd[1378]: cali3aa94ed9d93: Link UP Mar 7 01:50:19.777586 systemd-networkd[1378]: cali3aa94ed9d93: Gained carrier Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.704 [INFO][5152] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0 goldmane-5b85766d88- calico-system e48198d3-aa71-41df-a131-1b93c5c926c5 1000 0 2026-03-07 01:49:45 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:5b85766d88 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af goldmane-5b85766d88-58627 eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali3aa94ed9d93 [] [] }} ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.704 [INFO][5152] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.728 [INFO][5164] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" HandleID="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.737 [INFO][5164] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" HandleID="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000273350), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"goldmane-5b85766d88-58627", "timestamp":"2026-03-07 01:50:19.728477354 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400030b080)} Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.737 [INFO][5164] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.737 [INFO][5164] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.737 [INFO][5164] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.739 [INFO][5164] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.745 [INFO][5164] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.749 [INFO][5164] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.751 [INFO][5164] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.753 [INFO][5164] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.753 [INFO][5164] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.754 [INFO][5164] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.763 [INFO][5164] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.771 [INFO][5164] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.67/26] block=192.168.112.64/26 handle="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.771 [INFO][5164] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.67/26] handle="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.772 [INFO][5164] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:19.799738 containerd[1817]: 2026-03-07 01:50:19.772 [INFO][5164] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.67/26] IPv6=[] ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" HandleID="k8s-pod-network.4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.800916 containerd[1817]: 2026-03-07 01:50:19.774 [INFO][5152] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"e48198d3-aa71-41df-a131-1b93c5c926c5", ResourceVersion:"1000", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"goldmane-5b85766d88-58627", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.112.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3aa94ed9d93", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:19.800916 containerd[1817]: 2026-03-07 01:50:19.774 [INFO][5152] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.67/32] ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.800916 containerd[1817]: 2026-03-07 01:50:19.774 [INFO][5152] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3aa94ed9d93 ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.800916 containerd[1817]: 2026-03-07 01:50:19.777 [INFO][5152] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.800916 containerd[1817]: 2026-03-07 01:50:19.778 [INFO][5152] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"e48198d3-aa71-41df-a131-1b93c5c926c5", ResourceVersion:"1000", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca", Pod:"goldmane-5b85766d88-58627", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.112.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3aa94ed9d93", MAC:"86:96:40:f8:e5:fe", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:19.800916 containerd[1817]: 2026-03-07 01:50:19.795 [INFO][5152] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca" Namespace="calico-system" Pod="goldmane-5b85766d88-58627" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:19.825010 containerd[1817]: time="2026-03-07T01:50:19.824927893Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:19.825173 containerd[1817]: time="2026-03-07T01:50:19.824992053Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:19.825173 containerd[1817]: time="2026-03-07T01:50:19.825008973Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:19.825173 containerd[1817]: time="2026-03-07T01:50:19.825090333Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:19.873741 containerd[1817]: time="2026-03-07T01:50:19.873702944Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-5b85766d88-58627,Uid:e48198d3-aa71-41df-a131-1b93c5c926c5,Namespace:calico-system,Attempt:1,} returns sandbox id \"4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca\"" Mar 7 01:50:19.875159 containerd[1817]: time="2026-03-07T01:50:19.874997386Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\"" Mar 7 01:50:20.634128 systemd[1]: run-containerd-runc-k8s.io-4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca-runc.SDE8rF.mount: Deactivated successfully. Mar 7 01:50:21.504711 systemd-networkd[1378]: cali3aa94ed9d93: Gained IPv6LL Mar 7 01:50:21.549484 containerd[1817]: time="2026-03-07T01:50:21.549127422Z" level=info msg="StopPodSandbox for \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\"" Mar 7 01:50:21.550391 containerd[1817]: time="2026-03-07T01:50:21.549767104Z" level=info msg="StopPodSandbox for \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\"" Mar 7 01:50:21.552543 containerd[1817]: time="2026-03-07T01:50:21.552410389Z" level=info msg="StopPodSandbox for \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\"" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.634 [INFO][5278] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.635 [INFO][5278] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" iface="eth0" netns="/var/run/netns/cni-8faf9890-9510-d5c2-84f1-8a2a3f0b8ae1" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.635 [INFO][5278] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" iface="eth0" netns="/var/run/netns/cni-8faf9890-9510-d5c2-84f1-8a2a3f0b8ae1" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.635 [INFO][5278] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" iface="eth0" netns="/var/run/netns/cni-8faf9890-9510-d5c2-84f1-8a2a3f0b8ae1" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.635 [INFO][5278] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.635 [INFO][5278] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.695 [INFO][5293] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.696 [INFO][5293] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.696 [INFO][5293] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.718 [WARNING][5293] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.723 [INFO][5293] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.732 [INFO][5293] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:21.744645 containerd[1817]: 2026-03-07 01:50:21.734 [INFO][5278] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:21.750307 systemd[1]: run-netns-cni\x2d8faf9890\x2d9510\x2dd5c2\x2d84f1\x2d8a2a3f0b8ae1.mount: Deactivated successfully. Mar 7 01:50:21.758686 containerd[1817]: time="2026-03-07T01:50:21.758164083Z" level=info msg="TearDown network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\" successfully" Mar 7 01:50:21.758686 containerd[1817]: time="2026-03-07T01:50:21.758527764Z" level=info msg="StopPodSandbox for \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\" returns successfully" Mar 7 01:50:21.764370 containerd[1817]: time="2026-03-07T01:50:21.763895015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-gb6nc,Uid:f64343b8-a2cd-4274-908a-c1805396f8d4,Namespace:calico-system,Attempt:1,}" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.651 [INFO][5265] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.651 [INFO][5265] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" iface="eth0" netns="/var/run/netns/cni-8e969ef8-81d4-2b15-d276-c397e28f2544" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.651 [INFO][5265] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" iface="eth0" netns="/var/run/netns/cni-8e969ef8-81d4-2b15-d276-c397e28f2544" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.652 [INFO][5265] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" iface="eth0" netns="/var/run/netns/cni-8e969ef8-81d4-2b15-d276-c397e28f2544" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.652 [INFO][5265] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.652 [INFO][5265] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.722 [INFO][5299] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.722 [INFO][5299] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.733 [INFO][5299] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.744 [WARNING][5299] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.744 [INFO][5299] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.746 [INFO][5299] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:21.772456 containerd[1817]: 2026-03-07 01:50:21.756 [INFO][5265] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:21.774106 containerd[1817]: time="2026-03-07T01:50:21.773979435Z" level=info msg="TearDown network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\" successfully" Mar 7 01:50:21.774106 containerd[1817]: time="2026-03-07T01:50:21.774007195Z" level=info msg="StopPodSandbox for \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\" returns successfully" Mar 7 01:50:21.774166 systemd[1]: run-netns-cni\x2d8e969ef8\x2d81d4\x2d2b15\x2dd276\x2dc397e28f2544.mount: Deactivated successfully. Mar 7 01:50:21.776318 containerd[1817]: time="2026-03-07T01:50:21.776193560Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-qfvhc,Uid:58c58ebb-a76c-40e3-bec1-27853dcf0611,Namespace:calico-system,Attempt:1,}" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.659 [INFO][5279] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.659 [INFO][5279] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" iface="eth0" netns="/var/run/netns/cni-2f5fe665-96fc-962d-c360-59fa38ee3323" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.659 [INFO][5279] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" iface="eth0" netns="/var/run/netns/cni-2f5fe665-96fc-962d-c360-59fa38ee3323" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.660 [INFO][5279] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" iface="eth0" netns="/var/run/netns/cni-2f5fe665-96fc-962d-c360-59fa38ee3323" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.661 [INFO][5279] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.661 [INFO][5279] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.746 [INFO][5304] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.748 [INFO][5304] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.748 [INFO][5304] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.768 [WARNING][5304] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.768 [INFO][5304] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.773 [INFO][5304] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:21.783803 containerd[1817]: 2026-03-07 01:50:21.779 [INFO][5279] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:21.785825 containerd[1817]: time="2026-03-07T01:50:21.785782979Z" level=info msg="TearDown network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\" successfully" Mar 7 01:50:21.785825 containerd[1817]: time="2026-03-07T01:50:21.785818019Z" level=info msg="StopPodSandbox for \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\" returns successfully" Mar 7 01:50:21.787917 containerd[1817]: time="2026-03-07T01:50:21.787676823Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-68c448d44-wz8jw,Uid:8de81ec4-1390-4c8d-8e83-a56e0632cd6d,Namespace:calico-system,Attempt:1,}" Mar 7 01:50:21.788325 systemd[1]: run-netns-cni\x2d2f5fe665\x2d96fc\x2d962d\x2dc360\x2d59fa38ee3323.mount: Deactivated successfully. Mar 7 01:50:22.212577 systemd-networkd[1378]: cali8ebdb3bace5: Link UP Mar 7 01:50:22.221423 systemd-networkd[1378]: cali8ebdb3bace5: Gained carrier Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:21.997 [INFO][5319] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0 calico-apiserver-57bc98fcf9- calico-system f64343b8-a2cd-4274-908a-c1805396f8d4 1012 0 2026-03-07 01:49:45 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:57bc98fcf9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af calico-apiserver-57bc98fcf9-gb6nc eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali8ebdb3bace5 [] [] }} ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:21.997 [INFO][5319] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.078 [INFO][5354] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" HandleID="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.095 [INFO][5354] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" HandleID="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004dea0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"calico-apiserver-57bc98fcf9-gb6nc", "timestamp":"2026-03-07 01:50:22.078186728 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000184580)} Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.095 [INFO][5354] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.096 [INFO][5354] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.097 [INFO][5354] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.103 [INFO][5354] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.110 [INFO][5354] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.117 [INFO][5354] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.119 [INFO][5354] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.122 [INFO][5354] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.123 [INFO][5354] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.125 [INFO][5354] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459 Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.144 [INFO][5354] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.175 [INFO][5354] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.68/26] block=192.168.112.64/26 handle="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.176 [INFO][5354] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.68/26] handle="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.176 [INFO][5354] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:22.256492 containerd[1817]: 2026-03-07 01:50:22.176 [INFO][5354] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.68/26] IPv6=[] ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" HandleID="k8s-pod-network.48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:22.257061 containerd[1817]: 2026-03-07 01:50:22.185 [INFO][5319] cni-plugin/k8s.go 418: Populated endpoint ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"f64343b8-a2cd-4274-908a-c1805396f8d4", ResourceVersion:"1012", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"calico-apiserver-57bc98fcf9-gb6nc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali8ebdb3bace5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:22.257061 containerd[1817]: 2026-03-07 01:50:22.185 [INFO][5319] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.68/32] ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:22.257061 containerd[1817]: 2026-03-07 01:50:22.185 [INFO][5319] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8ebdb3bace5 ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:22.257061 containerd[1817]: 2026-03-07 01:50:22.224 [INFO][5319] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:22.257061 containerd[1817]: 2026-03-07 01:50:22.225 [INFO][5319] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"f64343b8-a2cd-4274-908a-c1805396f8d4", ResourceVersion:"1012", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459", Pod:"calico-apiserver-57bc98fcf9-gb6nc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali8ebdb3bace5", MAC:"8a:31:e6:5e:44:39", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:22.257061 containerd[1817]: 2026-03-07 01:50:22.247 [INFO][5319] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-gb6nc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:22.306139 systemd-networkd[1378]: calif8111e55bfe: Link UP Mar 7 01:50:22.307926 systemd-networkd[1378]: calif8111e55bfe: Gained carrier Mar 7 01:50:22.339961 containerd[1817]: time="2026-03-07T01:50:22.338584373Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:22.339961 containerd[1817]: time="2026-03-07T01:50:22.338647373Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:22.339961 containerd[1817]: time="2026-03-07T01:50:22.338668013Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:22.339961 containerd[1817]: time="2026-03-07T01:50:22.338758613Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.068 [INFO][5328] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0 calico-kube-controllers-68c448d44- calico-system 8de81ec4-1390-4c8d-8e83-a56e0632cd6d 1014 0 2026-03-07 01:49:47 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:68c448d44 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af calico-kube-controllers-68c448d44-wz8jw eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calif8111e55bfe [] [] }} ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.068 [INFO][5328] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.125 [INFO][5365] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" HandleID="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.139 [INFO][5365] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" HandleID="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fbaf0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"calico-kube-controllers-68c448d44-wz8jw", "timestamp":"2026-03-07 01:50:22.125713824 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400025ef20)} Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.139 [INFO][5365] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.176 [INFO][5365] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.176 [INFO][5365] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.208 [INFO][5365] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.248 [INFO][5365] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.255 [INFO][5365] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.258 [INFO][5365] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.261 [INFO][5365] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.261 [INFO][5365] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.264 [INFO][5365] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8 Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.279 [INFO][5365] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.289 [INFO][5365] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.69/26] block=192.168.112.64/26 handle="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.289 [INFO][5365] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.69/26] handle="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.290 [INFO][5365] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:22.357495 containerd[1817]: 2026-03-07 01:50:22.290 [INFO][5365] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.69/26] IPv6=[] ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" HandleID="k8s-pod-network.3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:22.358014 containerd[1817]: 2026-03-07 01:50:22.294 [INFO][5328] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0", GenerateName:"calico-kube-controllers-68c448d44-", Namespace:"calico-system", SelfLink:"", UID:"8de81ec4-1390-4c8d-8e83-a56e0632cd6d", ResourceVersion:"1014", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"68c448d44", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"calico-kube-controllers-68c448d44-wz8jw", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.112.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif8111e55bfe", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:22.358014 containerd[1817]: 2026-03-07 01:50:22.295 [INFO][5328] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.69/32] ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:22.358014 containerd[1817]: 2026-03-07 01:50:22.295 [INFO][5328] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif8111e55bfe ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:22.358014 containerd[1817]: 2026-03-07 01:50:22.311 [INFO][5328] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:22.358014 containerd[1817]: 2026-03-07 01:50:22.312 [INFO][5328] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0", GenerateName:"calico-kube-controllers-68c448d44-", Namespace:"calico-system", SelfLink:"", UID:"8de81ec4-1390-4c8d-8e83-a56e0632cd6d", ResourceVersion:"1014", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"68c448d44", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8", Pod:"calico-kube-controllers-68c448d44-wz8jw", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.112.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif8111e55bfe", MAC:"06:96:06:ca:36:58", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:22.358014 containerd[1817]: 2026-03-07 01:50:22.349 [INFO][5328] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8" Namespace="calico-system" Pod="calico-kube-controllers-68c448d44-wz8jw" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:22.431945 containerd[1817]: time="2026-03-07T01:50:22.431224479Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:22.431945 containerd[1817]: time="2026-03-07T01:50:22.431291599Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:22.431945 containerd[1817]: time="2026-03-07T01:50:22.431303279Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:22.431945 containerd[1817]: time="2026-03-07T01:50:22.431398640Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:22.438488 systemd-networkd[1378]: cali15066988c01: Link UP Mar 7 01:50:22.438990 systemd-networkd[1378]: cali15066988c01: Gained carrier Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.092 [INFO][5339] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0 calico-apiserver-57bc98fcf9- calico-system 58c58ebb-a76c-40e3-bec1-27853dcf0611 1013 0 2026-03-07 01:49:45 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:57bc98fcf9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af calico-apiserver-57bc98fcf9-qfvhc eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali15066988c01 [] [] }} ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.093 [INFO][5339] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.271 [INFO][5371] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" HandleID="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.297 [INFO][5371] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" HandleID="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fb440), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"calico-apiserver-57bc98fcf9-qfvhc", "timestamp":"2026-03-07 01:50:22.271948598 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x40000f2420)} Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.301 [INFO][5371] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.302 [INFO][5371] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.302 [INFO][5371] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.309 [INFO][5371] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.348 [INFO][5371] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.370 [INFO][5371] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.380 [INFO][5371] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.387 [INFO][5371] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.387 [INFO][5371] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.390 [INFO][5371] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01 Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.397 [INFO][5371] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.410 [INFO][5371] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.70/26] block=192.168.112.64/26 handle="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.410 [INFO][5371] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.70/26] handle="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.410 [INFO][5371] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:22.473646 containerd[1817]: 2026-03-07 01:50:22.410 [INFO][5371] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.70/26] IPv6=[] ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" HandleID="k8s-pod-network.dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:22.474154 containerd[1817]: 2026-03-07 01:50:22.419 [INFO][5339] cni-plugin/k8s.go 418: Populated endpoint ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"58c58ebb-a76c-40e3-bec1-27853dcf0611", ResourceVersion:"1013", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"calico-apiserver-57bc98fcf9-qfvhc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali15066988c01", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:22.474154 containerd[1817]: 2026-03-07 01:50:22.419 [INFO][5339] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.70/32] ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:22.474154 containerd[1817]: 2026-03-07 01:50:22.419 [INFO][5339] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali15066988c01 ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:22.474154 containerd[1817]: 2026-03-07 01:50:22.444 [INFO][5339] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:22.474154 containerd[1817]: 2026-03-07 01:50:22.449 [INFO][5339] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"58c58ebb-a76c-40e3-bec1-27853dcf0611", ResourceVersion:"1013", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01", Pod:"calico-apiserver-57bc98fcf9-qfvhc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali15066988c01", MAC:"da:74:73:87:48:0d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:22.474154 containerd[1817]: 2026-03-07 01:50:22.469 [INFO][5339] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01" Namespace="calico-system" Pod="calico-apiserver-57bc98fcf9-qfvhc" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:22.488022 containerd[1817]: time="2026-03-07T01:50:22.487658433Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-gb6nc,Uid:f64343b8-a2cd-4274-908a-c1805396f8d4,Namespace:calico-system,Attempt:1,} returns sandbox id \"48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459\"" Mar 7 01:50:22.522634 containerd[1817]: time="2026-03-07T01:50:22.522531463Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-68c448d44-wz8jw,Uid:8de81ec4-1390-4c8d-8e83-a56e0632cd6d,Namespace:calico-system,Attempt:1,} returns sandbox id \"3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8\"" Mar 7 01:50:22.532480 containerd[1817]: time="2026-03-07T01:50:22.528688876Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:22.532480 containerd[1817]: time="2026-03-07T01:50:22.528787636Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:22.532480 containerd[1817]: time="2026-03-07T01:50:22.528802196Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:22.532480 containerd[1817]: time="2026-03-07T01:50:22.528918076Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:22.620319 containerd[1817]: time="2026-03-07T01:50:22.620210980Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57bc98fcf9-qfvhc,Uid:58c58ebb-a76c-40e3-bec1-27853dcf0611,Namespace:calico-system,Attempt:1,} returns sandbox id \"dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01\"" Mar 7 01:50:22.689264 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3755589232.mount: Deactivated successfully. Mar 7 01:50:22.914416 containerd[1817]: time="2026-03-07T01:50:22.913901492Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:22.916726 containerd[1817]: time="2026-03-07T01:50:22.916700497Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.4: active requests=0, bytes read=51613980" Mar 7 01:50:22.919814 containerd[1817]: time="2026-03-07T01:50:22.919792024Z" level=info msg="ImageCreate event name:\"sha256:5274e98e9b12badfa0d6f106814630212e6de7abb8deaf896423b13e6ebdb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:22.929414 containerd[1817]: time="2026-03-07T01:50:22.929361843Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:22.930330 containerd[1817]: time="2026-03-07T01:50:22.930302165Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" with image id \"sha256:5274e98e9b12badfa0d6f106814630212e6de7abb8deaf896423b13e6ebdb41b\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\", size \"51613826\" in 3.055273699s" Mar 7 01:50:22.930413 containerd[1817]: time="2026-03-07T01:50:22.930400845Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" returns image reference \"sha256:5274e98e9b12badfa0d6f106814630212e6de7abb8deaf896423b13e6ebdb41b\"" Mar 7 01:50:22.933536 containerd[1817]: time="2026-03-07T01:50:22.933450931Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Mar 7 01:50:22.939361 containerd[1817]: time="2026-03-07T01:50:22.939160103Z" level=info msg="CreateContainer within sandbox \"4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Mar 7 01:50:22.981268 containerd[1817]: time="2026-03-07T01:50:22.981192067Z" level=info msg="CreateContainer within sandbox \"4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"5f69b74da65abbf05650c281ff7170ca87de1782f2219a4ffc6f1c2c0bb4fdf9\"" Mar 7 01:50:22.982185 containerd[1817]: time="2026-03-07T01:50:22.981915749Z" level=info msg="StartContainer for \"5f69b74da65abbf05650c281ff7170ca87de1782f2219a4ffc6f1c2c0bb4fdf9\"" Mar 7 01:50:23.047823 containerd[1817]: time="2026-03-07T01:50:23.047779841Z" level=info msg="StartContainer for \"5f69b74da65abbf05650c281ff7170ca87de1782f2219a4ffc6f1c2c0bb4fdf9\" returns successfully" Mar 7 01:50:23.296775 systemd-networkd[1378]: cali8ebdb3bace5: Gained IPv6LL Mar 7 01:50:23.360667 systemd-networkd[1378]: calif8111e55bfe: Gained IPv6LL Mar 7 01:50:23.549074 containerd[1817]: time="2026-03-07T01:50:23.548017329Z" level=info msg="StopPodSandbox for \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\"" Mar 7 01:50:23.549074 containerd[1817]: time="2026-03-07T01:50:23.548017209Z" level=info msg="StopPodSandbox for \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\"" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.609 [INFO][5627] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.610 [INFO][5627] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" iface="eth0" netns="/var/run/netns/cni-53d5d028-0a34-d942-4ae6-74deb311a766" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.610 [INFO][5627] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" iface="eth0" netns="/var/run/netns/cni-53d5d028-0a34-d942-4ae6-74deb311a766" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.611 [INFO][5627] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" iface="eth0" netns="/var/run/netns/cni-53d5d028-0a34-d942-4ae6-74deb311a766" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.611 [INFO][5627] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.611 [INFO][5627] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.648 [INFO][5640] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.648 [INFO][5640] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.648 [INFO][5640] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.658 [WARNING][5640] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.658 [INFO][5640] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.659 [INFO][5640] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:23.667647 containerd[1817]: 2026-03-07 01:50:23.662 [INFO][5627] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:23.669754 containerd[1817]: time="2026-03-07T01:50:23.669683374Z" level=info msg="TearDown network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\" successfully" Mar 7 01:50:23.670022 containerd[1817]: time="2026-03-07T01:50:23.669856055Z" level=info msg="StopPodSandbox for \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\" returns successfully" Mar 7 01:50:23.678726 containerd[1817]: time="2026-03-07T01:50:23.678694392Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n7zbg,Uid:8cd6b288-aa15-4bc8-b027-e2e3c62e23a5,Namespace:kube-system,Attempt:1,}" Mar 7 01:50:23.679257 systemd[1]: run-netns-cni\x2d53d5d028\x2d0a34\x2dd942\x2d4ae6\x2d74deb311a766.mount: Deactivated successfully. Mar 7 01:50:23.680558 systemd-networkd[1378]: cali15066988c01: Gained IPv6LL Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.613 [INFO][5626] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.614 [INFO][5626] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" iface="eth0" netns="/var/run/netns/cni-dc6c9cfe-f8d0-b5af-1b45-d9c097088904" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.616 [INFO][5626] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" iface="eth0" netns="/var/run/netns/cni-dc6c9cfe-f8d0-b5af-1b45-d9c097088904" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.616 [INFO][5626] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" iface="eth0" netns="/var/run/netns/cni-dc6c9cfe-f8d0-b5af-1b45-d9c097088904" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.616 [INFO][5626] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.616 [INFO][5626] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.651 [INFO][5645] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.651 [INFO][5645] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.659 [INFO][5645] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.670 [WARNING][5645] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.670 [INFO][5645] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.673 [INFO][5645] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:23.682565 containerd[1817]: 2026-03-07 01:50:23.676 [INFO][5626] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:23.683249 containerd[1817]: time="2026-03-07T01:50:23.682887281Z" level=info msg="TearDown network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\" successfully" Mar 7 01:50:23.683249 containerd[1817]: time="2026-03-07T01:50:23.682908401Z" level=info msg="StopPodSandbox for \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\" returns successfully" Mar 7 01:50:23.683417 containerd[1817]: time="2026-03-07T01:50:23.683382442Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hkgbg,Uid:a40182e6-4b51-42f9-8f1a-ad85b1365603,Namespace:kube-system,Attempt:1,}" Mar 7 01:50:23.687935 systemd[1]: run-netns-cni\x2ddc6c9cfe\x2df8d0\x2db5af\x2d1b45\x2dd9c097088904.mount: Deactivated successfully. Mar 7 01:50:23.931992 systemd-networkd[1378]: cali10648cd9737: Link UP Mar 7 01:50:23.933037 systemd-networkd[1378]: cali10648cd9737: Gained carrier Mar 7 01:50:23.947343 kubelet[3332]: I0307 01:50:23.947194 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-5b85766d88-58627" podStartSLOduration=35.890685752 podStartE2EDuration="38.947175413s" podCreationTimestamp="2026-03-07 01:49:45 +0000 UTC" firstStartedPulling="2026-03-07 01:50:19.874793626 +0000 UTC m=+51.422395662" lastFinishedPulling="2026-03-07 01:50:22.931283287 +0000 UTC m=+54.478885323" observedRunningTime="2026-03-07 01:50:23.815582868 +0000 UTC m=+55.363184904" watchObservedRunningTime="2026-03-07 01:50:23.947175413 +0000 UTC m=+55.494777409" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.795 [INFO][5658] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0 coredns-674b8bbfcf- kube-system a40182e6-4b51-42f9-8f1a-ad85b1365603 1036 0 2026-03-07 01:49:35 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af coredns-674b8bbfcf-hkgbg eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali10648cd9737 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.795 [INFO][5658] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.872 [INFO][5691] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" HandleID="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.891 [INFO][5691] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" HandleID="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40005fa350), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"coredns-674b8bbfcf-hkgbg", "timestamp":"2026-03-07 01:50:23.872851503 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x4000185080)} Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.891 [INFO][5691] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.891 [INFO][5691] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.891 [INFO][5691] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.896 [INFO][5691] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.902 [INFO][5691] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.905 [INFO][5691] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.907 [INFO][5691] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.909 [INFO][5691] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.909 [INFO][5691] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.910 [INFO][5691] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8 Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.917 [INFO][5691] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.926 [INFO][5691] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.71/26] block=192.168.112.64/26 handle="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.926 [INFO][5691] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.71/26] handle="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.926 [INFO][5691] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:23.949696 containerd[1817]: 2026-03-07 01:50:23.926 [INFO][5691] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.71/26] IPv6=[] ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" HandleID="k8s-pod-network.3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.950171 containerd[1817]: 2026-03-07 01:50:23.929 [INFO][5658] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a40182e6-4b51-42f9-8f1a-ad85b1365603", ResourceVersion:"1036", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"coredns-674b8bbfcf-hkgbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali10648cd9737", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:23.950171 containerd[1817]: 2026-03-07 01:50:23.929 [INFO][5658] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.71/32] ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.950171 containerd[1817]: 2026-03-07 01:50:23.929 [INFO][5658] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali10648cd9737 ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.950171 containerd[1817]: 2026-03-07 01:50:23.933 [INFO][5658] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.950171 containerd[1817]: 2026-03-07 01:50:23.933 [INFO][5658] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a40182e6-4b51-42f9-8f1a-ad85b1365603", ResourceVersion:"1036", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8", Pod:"coredns-674b8bbfcf-hkgbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali10648cd9737", MAC:"22:ee:7a:be:f6:c2", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:23.950171 containerd[1817]: 2026-03-07 01:50:23.946 [INFO][5658] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8" Namespace="kube-system" Pod="coredns-674b8bbfcf-hkgbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:23.979054 containerd[1817]: time="2026-03-07T01:50:23.977719035Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:23.979054 containerd[1817]: time="2026-03-07T01:50:23.977864235Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:23.979054 containerd[1817]: time="2026-03-07T01:50:23.977903435Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:23.979054 containerd[1817]: time="2026-03-07T01:50:23.978095276Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:24.057109 containerd[1817]: time="2026-03-07T01:50:24.057063555Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hkgbg,Uid:a40182e6-4b51-42f9-8f1a-ad85b1365603,Namespace:kube-system,Attempt:1,} returns sandbox id \"3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8\"" Mar 7 01:50:24.071349 containerd[1817]: time="2026-03-07T01:50:24.070536062Z" level=info msg="CreateContainer within sandbox \"3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 7 01:50:24.076888 systemd-networkd[1378]: cali4a1cdb78ae7: Link UP Mar 7 01:50:24.077992 systemd-networkd[1378]: cali4a1cdb78ae7: Gained carrier Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.826 [INFO][5673] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0 coredns-674b8bbfcf- kube-system 8cd6b288-aa15-4bc8-b027-e2e3c62e23a5 1035 0 2026-03-07 01:49:35 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.6-n-e8b1fc37af coredns-674b8bbfcf-n7zbg eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali4a1cdb78ae7 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.826 [INFO][5673] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.884 [INFO][5704] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" HandleID="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.898 [INFO][5704] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" HandleID="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003ea140), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.6-n-e8b1fc37af", "pod":"coredns-674b8bbfcf-n7zbg", "timestamp":"2026-03-07 01:50:23.884125486 +0000 UTC"}, Hostname:"ci-4081.3.6-n-e8b1fc37af", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0x400010c2c0)} Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.898 [INFO][5704] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.926 [INFO][5704] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.927 [INFO][5704] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.6-n-e8b1fc37af' Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:23.996 [INFO][5704] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.010 [INFO][5704] ipam/ipam.go 409: Looking up existing affinities for host host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.027 [INFO][5704] ipam/ipam.go 526: Trying affinity for 192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.031 [INFO][5704] ipam/ipam.go 160: Attempting to load block cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.036 [INFO][5704] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.112.64/26 host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.036 [INFO][5704] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.112.64/26 handle="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.038 [INFO][5704] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.047 [INFO][5704] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.112.64/26 handle="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.061 [INFO][5704] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.112.72/26] block=192.168.112.64/26 handle="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.062 [INFO][5704] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.112.72/26] handle="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" host="ci-4081.3.6-n-e8b1fc37af" Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.062 [INFO][5704] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:24.098678 containerd[1817]: 2026-03-07 01:50:24.062 [INFO][5704] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.112.72/26] IPv6=[] ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" HandleID="k8s-pod-network.cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:24.099740 containerd[1817]: 2026-03-07 01:50:24.065 [INFO][5673] cni-plugin/k8s.go 418: Populated endpoint ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5", ResourceVersion:"1035", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"", Pod:"coredns-674b8bbfcf-n7zbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4a1cdb78ae7", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:24.099740 containerd[1817]: 2026-03-07 01:50:24.067 [INFO][5673] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.112.72/32] ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:24.099740 containerd[1817]: 2026-03-07 01:50:24.067 [INFO][5673] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4a1cdb78ae7 ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:24.099740 containerd[1817]: 2026-03-07 01:50:24.078 [INFO][5673] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:24.099740 containerd[1817]: 2026-03-07 01:50:24.079 [INFO][5673] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5", ResourceVersion:"1035", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e", Pod:"coredns-674b8bbfcf-n7zbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4a1cdb78ae7", MAC:"26:30:6b:9e:88:f1", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:24.099740 containerd[1817]: 2026-03-07 01:50:24.095 [INFO][5673] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e" Namespace="kube-system" Pod="coredns-674b8bbfcf-n7zbg" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:24.129470 containerd[1817]: time="2026-03-07T01:50:24.128500859Z" level=info msg="CreateContainer within sandbox \"3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9e6df933f6171acb29fcfddb77fe1a71c4345ddb0c48a092d495ecc4d081b6c5\"" Mar 7 01:50:24.131205 containerd[1817]: time="2026-03-07T01:50:24.131176224Z" level=info msg="StartContainer for \"9e6df933f6171acb29fcfddb77fe1a71c4345ddb0c48a092d495ecc4d081b6c5\"" Mar 7 01:50:24.134202 containerd[1817]: time="2026-03-07T01:50:24.134119590Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 01:50:24.134329 containerd[1817]: time="2026-03-07T01:50:24.134266590Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 01:50:24.134329 containerd[1817]: time="2026-03-07T01:50:24.134306870Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:24.134571 containerd[1817]: time="2026-03-07T01:50:24.134531031Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 01:50:24.203619 containerd[1817]: time="2026-03-07T01:50:24.202738288Z" level=info msg="StartContainer for \"9e6df933f6171acb29fcfddb77fe1a71c4345ddb0c48a092d495ecc4d081b6c5\" returns successfully" Mar 7 01:50:24.204233 containerd[1817]: time="2026-03-07T01:50:24.204210331Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n7zbg,Uid:8cd6b288-aa15-4bc8-b027-e2e3c62e23a5,Namespace:kube-system,Attempt:1,} returns sandbox id \"cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e\"" Mar 7 01:50:24.233789 containerd[1817]: time="2026-03-07T01:50:24.233632350Z" level=info msg="CreateContainer within sandbox \"cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 7 01:50:24.271816 containerd[1817]: time="2026-03-07T01:50:24.271771747Z" level=info msg="CreateContainer within sandbox \"cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"01071f0c8814468f2c89f8db4f5f69881165ddcf0c2f55e82e9dc735d4a61382\"" Mar 7 01:50:24.272734 containerd[1817]: time="2026-03-07T01:50:24.272702189Z" level=info msg="StartContainer for \"01071f0c8814468f2c89f8db4f5f69881165ddcf0c2f55e82e9dc735d4a61382\"" Mar 7 01:50:24.329449 containerd[1817]: time="2026-03-07T01:50:24.329394943Z" level=info msg="StartContainer for \"01071f0c8814468f2c89f8db4f5f69881165ddcf0c2f55e82e9dc735d4a61382\" returns successfully" Mar 7 01:50:24.844176 kubelet[3332]: I0307 01:50:24.843229 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-n7zbg" podStartSLOduration=49.843212978 podStartE2EDuration="49.843212978s" podCreationTimestamp="2026-03-07 01:49:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:50:24.813876279 +0000 UTC m=+56.361478355" watchObservedRunningTime="2026-03-07 01:50:24.843212978 +0000 UTC m=+56.390815014" Mar 7 01:50:24.867256 kubelet[3332]: I0307 01:50:24.867200 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-hkgbg" podStartSLOduration=49.867184747 podStartE2EDuration="49.867184747s" podCreationTimestamp="2026-03-07 01:49:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 01:50:24.844668341 +0000 UTC m=+56.392270417" watchObservedRunningTime="2026-03-07 01:50:24.867184747 +0000 UTC m=+56.414786783" Mar 7 01:50:25.024586 systemd-networkd[1378]: cali10648cd9737: Gained IPv6LL Mar 7 01:50:26.048596 systemd-networkd[1378]: cali4a1cdb78ae7: Gained IPv6LL Mar 7 01:50:26.258762 containerd[1817]: time="2026-03-07T01:50:26.257889068Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:26.261848 containerd[1817]: time="2026-03-07T01:50:26.261813196Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=45552315" Mar 7 01:50:26.265262 containerd[1817]: time="2026-03-07T01:50:26.265201683Z" level=info msg="ImageCreate event name:\"sha256:dca640051f09574f3e8821035bbfae8c638fb7dadca4c9a082e7223a234befc8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:26.270467 containerd[1817]: time="2026-03-07T01:50:26.269984493Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:26.270827 containerd[1817]: time="2026-03-07T01:50:26.270680534Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:dca640051f09574f3e8821035bbfae8c638fb7dadca4c9a082e7223a234befc8\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"46949856\" in 3.337192563s" Mar 7 01:50:26.270827 containerd[1817]: time="2026-03-07T01:50:26.270712654Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:dca640051f09574f3e8821035bbfae8c638fb7dadca4c9a082e7223a234befc8\"" Mar 7 01:50:26.272017 containerd[1817]: time="2026-03-07T01:50:26.271988217Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\"" Mar 7 01:50:26.290966 containerd[1817]: time="2026-03-07T01:50:26.290916815Z" level=info msg="CreateContainer within sandbox \"48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 7 01:50:26.330664 containerd[1817]: time="2026-03-07T01:50:26.330418934Z" level=info msg="CreateContainer within sandbox \"48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"c1d7c3d795e75d6ff4d255be04248481dd41a4cb5251870dbfc29a4a89d69199\"" Mar 7 01:50:26.332498 containerd[1817]: time="2026-03-07T01:50:26.332405418Z" level=info msg="StartContainer for \"c1d7c3d795e75d6ff4d255be04248481dd41a4cb5251870dbfc29a4a89d69199\"" Mar 7 01:50:26.415301 containerd[1817]: time="2026-03-07T01:50:26.415248945Z" level=info msg="StartContainer for \"c1d7c3d795e75d6ff4d255be04248481dd41a4cb5251870dbfc29a4a89d69199\" returns successfully" Mar 7 01:50:26.822099 kubelet[3332]: I0307 01:50:26.821951 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-57bc98fcf9-gb6nc" podStartSLOduration=38.041263629 podStartE2EDuration="41.821932405s" podCreationTimestamp="2026-03-07 01:49:45 +0000 UTC" firstStartedPulling="2026-03-07 01:50:22.49119552 +0000 UTC m=+54.038797516" lastFinishedPulling="2026-03-07 01:50:26.271864296 +0000 UTC m=+57.819466292" observedRunningTime="2026-03-07 01:50:26.820262481 +0000 UTC m=+58.367864517" watchObservedRunningTime="2026-03-07 01:50:26.821932405 +0000 UTC m=+58.369534441" Mar 7 01:50:27.809210 kubelet[3332]: I0307 01:50:27.809146 3332 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:50:28.551516 containerd[1817]: time="2026-03-07T01:50:28.551208008Z" level=info msg="StopPodSandbox for \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\"" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.594 [WARNING][6030] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"e48198d3-aa71-41df-a131-1b93c5c926c5", ResourceVersion:"1038", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca", Pod:"goldmane-5b85766d88-58627", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.112.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3aa94ed9d93", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.594 [INFO][6030] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.594 [INFO][6030] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" iface="eth0" netns="" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.594 [INFO][6030] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.594 [INFO][6030] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.629 [INFO][6037] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.629 [INFO][6037] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.630 [INFO][6037] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.638 [WARNING][6037] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.638 [INFO][6037] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.640 [INFO][6037] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:28.644759 containerd[1817]: 2026-03-07 01:50:28.642 [INFO][6030] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.645544 containerd[1817]: time="2026-03-07T01:50:28.645222318Z" level=info msg="TearDown network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\" successfully" Mar 7 01:50:28.645544 containerd[1817]: time="2026-03-07T01:50:28.645250358Z" level=info msg="StopPodSandbox for \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\" returns successfully" Mar 7 01:50:28.646024 containerd[1817]: time="2026-03-07T01:50:28.645974119Z" level=info msg="RemovePodSandbox for \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\"" Mar 7 01:50:28.646024 containerd[1817]: time="2026-03-07T01:50:28.646004599Z" level=info msg="Forcibly stopping sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\"" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.677 [WARNING][6051] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0", GenerateName:"goldmane-5b85766d88-", Namespace:"calico-system", SelfLink:"", UID:"e48198d3-aa71-41df-a131-1b93c5c926c5", ResourceVersion:"1038", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"5b85766d88", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"4789e6e20981d3fa1a72d825a6e474aa358d7b0f5b473276e40aa82c4917fdca", Pod:"goldmane-5b85766d88-58627", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.112.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3aa94ed9d93", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.677 [INFO][6051] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.677 [INFO][6051] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" iface="eth0" netns="" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.677 [INFO][6051] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.677 [INFO][6051] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.696 [INFO][6058] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.696 [INFO][6058] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.696 [INFO][6058] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.705 [WARNING][6058] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.705 [INFO][6058] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" HandleID="k8s-pod-network.619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-goldmane--5b85766d88--58627-eth0" Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.707 [INFO][6058] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:28.711556 containerd[1817]: 2026-03-07 01:50:28.708 [INFO][6051] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2" Mar 7 01:50:28.711556 containerd[1817]: time="2026-03-07T01:50:28.710597249Z" level=info msg="TearDown network for sandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\" successfully" Mar 7 01:50:28.870547 containerd[1817]: time="2026-03-07T01:50:28.870508159Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:50:28.870871 containerd[1817]: time="2026-03-07T01:50:28.870732439Z" level=info msg="RemovePodSandbox \"619e5a46320708a81e0f073f2a61c6c88264dfe8ceea66dd9ca681c4c5ef3fe2\" returns successfully" Mar 7 01:50:28.871297 containerd[1817]: time="2026-03-07T01:50:28.871263481Z" level=info msg="StopPodSandbox for \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\"" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.932 [WARNING][6076] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5", ResourceVersion:"1061", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e", Pod:"coredns-674b8bbfcf-n7zbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4a1cdb78ae7", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.932 [INFO][6076] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.932 [INFO][6076] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" iface="eth0" netns="" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.932 [INFO][6076] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.932 [INFO][6076] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.962 [INFO][6083] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.962 [INFO][6083] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.962 [INFO][6083] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.972 [WARNING][6083] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.972 [INFO][6083] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.973 [INFO][6083] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:28.979201 containerd[1817]: 2026-03-07 01:50:28.976 [INFO][6076] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:28.979948 containerd[1817]: time="2026-03-07T01:50:28.979921762Z" level=info msg="TearDown network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\" successfully" Mar 7 01:50:28.980005 containerd[1817]: time="2026-03-07T01:50:28.979992842Z" level=info msg="StopPodSandbox for \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\" returns successfully" Mar 7 01:50:28.980805 containerd[1817]: time="2026-03-07T01:50:28.980733524Z" level=info msg="RemovePodSandbox for \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\"" Mar 7 01:50:28.980873 containerd[1817]: time="2026-03-07T01:50:28.980811924Z" level=info msg="Forcibly stopping sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\"" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.029 [WARNING][6097] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"8cd6b288-aa15-4bc8-b027-e2e3c62e23a5", ResourceVersion:"1061", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"cd0acc26cf71d9ad7b862f0b46a99d905007ee612d427a99e6306aaef06bc65e", Pod:"coredns-674b8bbfcf-n7zbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4a1cdb78ae7", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.030 [INFO][6097] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.030 [INFO][6097] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" iface="eth0" netns="" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.030 [INFO][6097] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.030 [INFO][6097] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.061 [INFO][6104] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.062 [INFO][6104] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.062 [INFO][6104] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.072 [WARNING][6104] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.072 [INFO][6104] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" HandleID="k8s-pod-network.b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--n7zbg-eth0" Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.073 [INFO][6104] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:29.079465 containerd[1817]: 2026-03-07 01:50:29.075 [INFO][6097] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853" Mar 7 01:50:29.079465 containerd[1817]: time="2026-03-07T01:50:29.078633893Z" level=info msg="TearDown network for sandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\" successfully" Mar 7 01:50:29.088273 containerd[1817]: time="2026-03-07T01:50:29.088240521Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:50:29.088531 containerd[1817]: time="2026-03-07T01:50:29.088508242Z" level=info msg="RemovePodSandbox \"b62340a6cb5f87815dee3f940ca0e71444ce0aabbd4b7981a58de2df52b46853\" returns successfully" Mar 7 01:50:29.089102 containerd[1817]: time="2026-03-07T01:50:29.089080804Z" level=info msg="StopPodSandbox for \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\"" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.135 [WARNING][6118] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a40182e6-4b51-42f9-8f1a-ad85b1365603", ResourceVersion:"1069", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8", Pod:"coredns-674b8bbfcf-hkgbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali10648cd9737", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.135 [INFO][6118] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.135 [INFO][6118] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" iface="eth0" netns="" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.135 [INFO][6118] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.135 [INFO][6118] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.167 [INFO][6125] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.168 [INFO][6125] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.168 [INFO][6125] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.181 [WARNING][6125] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.181 [INFO][6125] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.182 [INFO][6125] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:29.194000 containerd[1817]: 2026-03-07 01:50:29.186 [INFO][6118] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.195586 containerd[1817]: time="2026-03-07T01:50:29.195517958Z" level=info msg="TearDown network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\" successfully" Mar 7 01:50:29.195586 containerd[1817]: time="2026-03-07T01:50:29.195547718Z" level=info msg="StopPodSandbox for \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\" returns successfully" Mar 7 01:50:29.198680 containerd[1817]: time="2026-03-07T01:50:29.198283966Z" level=info msg="RemovePodSandbox for \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\"" Mar 7 01:50:29.198680 containerd[1817]: time="2026-03-07T01:50:29.198330006Z" level=info msg="Forcibly stopping sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\"" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.239 [WARNING][6145] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a40182e6-4b51-42f9-8f1a-ad85b1365603", ResourceVersion:"1069", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"3ba5c8ccb87b4792eb8fdd73524fd283a773e152be0170ab2dab775ac1ec55a8", Pod:"coredns-674b8bbfcf-hkgbg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.112.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali10648cd9737", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.239 [INFO][6145] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.239 [INFO][6145] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" iface="eth0" netns="" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.239 [INFO][6145] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.239 [INFO][6145] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.265 [INFO][6153] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.265 [INFO][6153] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.265 [INFO][6153] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.273 [WARNING][6153] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.273 [INFO][6153] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" HandleID="k8s-pod-network.143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-coredns--674b8bbfcf--hkgbg-eth0" Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.275 [INFO][6153] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:29.281920 containerd[1817]: 2026-03-07 01:50:29.279 [INFO][6145] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5" Mar 7 01:50:29.282571 containerd[1817]: time="2026-03-07T01:50:29.282537854Z" level=info msg="TearDown network for sandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\" successfully" Mar 7 01:50:29.291163 containerd[1817]: time="2026-03-07T01:50:29.291115560Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:50:29.291423 containerd[1817]: time="2026-03-07T01:50:29.291388840Z" level=info msg="RemovePodSandbox \"143e68491c3cf6fed434914901791fce9406abf2c4382359a46b80b8cd2bbba5\" returns successfully" Mar 7 01:50:29.292652 containerd[1817]: time="2026-03-07T01:50:29.292589004Z" level=info msg="StopPodSandbox for \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\"" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.336 [WARNING][6168] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.336 [INFO][6168] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.336 [INFO][6168] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" iface="eth0" netns="" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.336 [INFO][6168] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.336 [INFO][6168] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.366 [INFO][6175] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.366 [INFO][6175] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.366 [INFO][6175] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.376 [WARNING][6175] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.376 [INFO][6175] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.377 [INFO][6175] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:29.381868 containerd[1817]: 2026-03-07 01:50:29.379 [INFO][6168] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.381868 containerd[1817]: time="2026-03-07T01:50:29.381755627Z" level=info msg="TearDown network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\" successfully" Mar 7 01:50:29.381868 containerd[1817]: time="2026-03-07T01:50:29.381781387Z" level=info msg="StopPodSandbox for \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\" returns successfully" Mar 7 01:50:29.382773 containerd[1817]: time="2026-03-07T01:50:29.382377029Z" level=info msg="RemovePodSandbox for \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\"" Mar 7 01:50:29.382773 containerd[1817]: time="2026-03-07T01:50:29.382502829Z" level=info msg="Forcibly stopping sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\"" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.430 [WARNING][6189] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" WorkloadEndpoint="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.430 [INFO][6189] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.430 [INFO][6189] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" iface="eth0" netns="" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.430 [INFO][6189] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.430 [INFO][6189] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.460 [INFO][6196] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.460 [INFO][6196] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.460 [INFO][6196] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.471 [WARNING][6196] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.472 [INFO][6196] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" HandleID="k8s-pod-network.ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-whisker--6b776dffb4--p5tkl-eth0" Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.473 [INFO][6196] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:29.477558 containerd[1817]: 2026-03-07 01:50:29.475 [INFO][6189] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8" Mar 7 01:50:29.478401 containerd[1817]: time="2026-03-07T01:50:29.477545270Z" level=info msg="TearDown network for sandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\" successfully" Mar 7 01:50:29.489116 containerd[1817]: time="2026-03-07T01:50:29.488977743Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:50:29.489650 containerd[1817]: time="2026-03-07T01:50:29.489606625Z" level=info msg="RemovePodSandbox \"ec0e0ae0e97e8bc8e14d1b336f0e4506972acf8dcec16c90cde259a2a1c22ea8\" returns successfully" Mar 7 01:50:29.491108 containerd[1817]: time="2026-03-07T01:50:29.490837789Z" level=info msg="StopPodSandbox for \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\"" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.535 [WARNING][6210] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"f64343b8-a2cd-4274-908a-c1805396f8d4", ResourceVersion:"1079", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459", Pod:"calico-apiserver-57bc98fcf9-gb6nc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali8ebdb3bace5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.535 [INFO][6210] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.535 [INFO][6210] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" iface="eth0" netns="" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.535 [INFO][6210] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.535 [INFO][6210] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.559 [INFO][6217] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.559 [INFO][6217] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.559 [INFO][6217] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.571 [WARNING][6217] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.571 [INFO][6217] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.573 [INFO][6217] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:29.579166 containerd[1817]: 2026-03-07 01:50:29.576 [INFO][6210] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.580503 containerd[1817]: time="2026-03-07T01:50:29.579202409Z" level=info msg="TearDown network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\" successfully" Mar 7 01:50:29.580503 containerd[1817]: time="2026-03-07T01:50:29.579226850Z" level=info msg="StopPodSandbox for \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\" returns successfully" Mar 7 01:50:29.580503 containerd[1817]: time="2026-03-07T01:50:29.579946892Z" level=info msg="RemovePodSandbox for \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\"" Mar 7 01:50:29.580503 containerd[1817]: time="2026-03-07T01:50:29.580045732Z" level=info msg="Forcibly stopping sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\"" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.625 [WARNING][6231] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"f64343b8-a2cd-4274-908a-c1805396f8d4", ResourceVersion:"1079", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"48ae6d278f8ab289b8ce789683cacf7ab742e79a581b745d218970f25deec459", Pod:"calico-apiserver-57bc98fcf9-gb6nc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali8ebdb3bace5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.625 [INFO][6231] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.625 [INFO][6231] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" iface="eth0" netns="" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.625 [INFO][6231] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.625 [INFO][6231] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.658 [INFO][6238] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.658 [INFO][6238] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.659 [INFO][6238] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.669 [WARNING][6238] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.669 [INFO][6238] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" HandleID="k8s-pod-network.a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--gb6nc-eth0" Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.671 [INFO][6238] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:29.676472 containerd[1817]: 2026-03-07 01:50:29.673 [INFO][6231] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e" Mar 7 01:50:29.676472 containerd[1817]: time="2026-03-07T01:50:29.675688534Z" level=info msg="TearDown network for sandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\" successfully" Mar 7 01:50:30.397078 containerd[1817]: time="2026-03-07T01:50:30.396730221Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:50:30.397078 containerd[1817]: time="2026-03-07T01:50:30.396814581Z" level=info msg="RemovePodSandbox \"a5f0a6e734fb4d92ed035fe596edacb12c725443443bfc7010964845068c2a4e\" returns successfully" Mar 7 01:50:30.398164 containerd[1817]: time="2026-03-07T01:50:30.398053905Z" level=info msg="StopPodSandbox for \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\"" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.440 [WARNING][6253] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0", GenerateName:"calico-kube-controllers-68c448d44-", Namespace:"calico-system", SelfLink:"", UID:"8de81ec4-1390-4c8d-8e83-a56e0632cd6d", ResourceVersion:"1022", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"68c448d44", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8", Pod:"calico-kube-controllers-68c448d44-wz8jw", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.112.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif8111e55bfe", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.440 [INFO][6253] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.440 [INFO][6253] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" iface="eth0" netns="" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.440 [INFO][6253] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.441 [INFO][6253] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.461 [INFO][6261] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.461 [INFO][6261] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.461 [INFO][6261] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.470 [WARNING][6261] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.470 [INFO][6261] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.471 [INFO][6261] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:30.474826 containerd[1817]: 2026-03-07 01:50:30.473 [INFO][6253] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.474826 containerd[1817]: time="2026-03-07T01:50:30.474710611Z" level=info msg="TearDown network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\" successfully" Mar 7 01:50:30.474826 containerd[1817]: time="2026-03-07T01:50:30.474735171Z" level=info msg="StopPodSandbox for \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\" returns successfully" Mar 7 01:50:30.475270 containerd[1817]: time="2026-03-07T01:50:30.475172533Z" level=info msg="RemovePodSandbox for \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\"" Mar 7 01:50:30.475270 containerd[1817]: time="2026-03-07T01:50:30.475199253Z" level=info msg="Forcibly stopping sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\"" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.506 [WARNING][6275] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0", GenerateName:"calico-kube-controllers-68c448d44-", Namespace:"calico-system", SelfLink:"", UID:"8de81ec4-1390-4c8d-8e83-a56e0632cd6d", ResourceVersion:"1022", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"68c448d44", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8", Pod:"calico-kube-controllers-68c448d44-wz8jw", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.112.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif8111e55bfe", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.506 [INFO][6275] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.506 [INFO][6275] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" iface="eth0" netns="" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.506 [INFO][6275] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.506 [INFO][6275] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.526 [INFO][6282] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.527 [INFO][6282] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.527 [INFO][6282] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.535 [WARNING][6282] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.535 [INFO][6282] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" HandleID="k8s-pod-network.f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--kube--controllers--68c448d44--wz8jw-eth0" Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.536 [INFO][6282] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:30.540002 containerd[1817]: 2026-03-07 01:50:30.538 [INFO][6275] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab" Mar 7 01:50:30.540401 containerd[1817]: time="2026-03-07T01:50:30.540042684Z" level=info msg="TearDown network for sandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\" successfully" Mar 7 01:50:30.973094 containerd[1817]: time="2026-03-07T01:50:30.973042041Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:50:30.973518 containerd[1817]: time="2026-03-07T01:50:30.973124522Z" level=info msg="RemovePodSandbox \"f44545aac4b7fcd601e5dbea16f3bc27421838357fdfdbc8730765807774beab\" returns successfully" Mar 7 01:50:30.974208 containerd[1817]: time="2026-03-07T01:50:30.973901284Z" level=info msg="StopPodSandbox for \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\"" Mar 7 01:50:30.992418 containerd[1817]: time="2026-03-07T01:50:30.992377938Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:30.993155 containerd[1817]: time="2026-03-07T01:50:30.993105621Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.4: active requests=0, bytes read=49189955" Mar 7 01:50:30.995770 containerd[1817]: time="2026-03-07T01:50:30.995736388Z" level=info msg="ImageCreate event name:\"sha256:e80fe1ce4f06b0791c077492cd9d5ebf00125a02bbafdcd04d2a64e10cc4ad95\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:31.006298 containerd[1817]: time="2026-03-07T01:50:31.005379977Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:31.010112 containerd[1817]: time="2026-03-07T01:50:31.010078871Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" with image id \"sha256:e80fe1ce4f06b0791c077492cd9d5ebf00125a02bbafdcd04d2a64e10cc4ad95\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\", size \"50587448\" in 4.738063614s" Mar 7 01:50:31.010191 containerd[1817]: time="2026-03-07T01:50:31.010114111Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" returns image reference \"sha256:e80fe1ce4f06b0791c077492cd9d5ebf00125a02bbafdcd04d2a64e10cc4ad95\"" Mar 7 01:50:31.013628 containerd[1817]: time="2026-03-07T01:50:31.013601601Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Mar 7 01:50:31.038124 containerd[1817]: time="2026-03-07T01:50:31.038090593Z" level=info msg="CreateContainer within sandbox \"3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.010 [WARNING][6297] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"58c58ebb-a76c-40e3-bec1-27853dcf0611", ResourceVersion:"1025", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01", Pod:"calico-apiserver-57bc98fcf9-qfvhc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali15066988c01", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.010 [INFO][6297] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.010 [INFO][6297] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" iface="eth0" netns="" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.010 [INFO][6297] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.010 [INFO][6297] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.048 [INFO][6305] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.048 [INFO][6305] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.048 [INFO][6305] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.059 [WARNING][6305] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.059 [INFO][6305] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.066 [INFO][6305] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:31.070553 containerd[1817]: 2026-03-07 01:50:31.068 [INFO][6297] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.073098 containerd[1817]: time="2026-03-07T01:50:31.070586329Z" level=info msg="TearDown network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\" successfully" Mar 7 01:50:31.073098 containerd[1817]: time="2026-03-07T01:50:31.070610129Z" level=info msg="StopPodSandbox for \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\" returns successfully" Mar 7 01:50:31.073098 containerd[1817]: time="2026-03-07T01:50:31.071041691Z" level=info msg="RemovePodSandbox for \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\"" Mar 7 01:50:31.073098 containerd[1817]: time="2026-03-07T01:50:31.071068011Z" level=info msg="Forcibly stopping sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\"" Mar 7 01:50:31.074762 containerd[1817]: time="2026-03-07T01:50:31.074661541Z" level=info msg="CreateContainer within sandbox \"3f5f4f2db60e799b77fba674b085bbff716a3ff24eac0ff6454f2b9bfefd22e8\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"471aa9f8cf9f70e5a5957091dbe6d6abe46a2ff89874d943b91ed33c0dc31e39\"" Mar 7 01:50:31.076562 containerd[1817]: time="2026-03-07T01:50:31.075339703Z" level=info msg="StartContainer for \"471aa9f8cf9f70e5a5957091dbe6d6abe46a2ff89874d943b91ed33c0dc31e39\"" Mar 7 01:50:31.153841 containerd[1817]: time="2026-03-07T01:50:31.153731374Z" level=info msg="StartContainer for \"471aa9f8cf9f70e5a5957091dbe6d6abe46a2ff89874d943b91ed33c0dc31e39\" returns successfully" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.127 [WARNING][6321] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0", GenerateName:"calico-apiserver-57bc98fcf9-", Namespace:"calico-system", SelfLink:"", UID:"58c58ebb-a76c-40e3-bec1-27853dcf0611", ResourceVersion:"1025", Generation:0, CreationTimestamp:time.Date(2026, time.March, 7, 1, 49, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57bc98fcf9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.6-n-e8b1fc37af", ContainerID:"dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01", Pod:"calico-apiserver-57bc98fcf9-qfvhc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.112.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali15066988c01", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.127 [INFO][6321] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.127 [INFO][6321] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" iface="eth0" netns="" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.127 [INFO][6321] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.127 [INFO][6321] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.157 [INFO][6352] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.157 [INFO][6352] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.157 [INFO][6352] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.172 [WARNING][6352] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.172 [INFO][6352] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" HandleID="k8s-pod-network.6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Workload="ci--4081.3.6--n--e8b1fc37af-k8s-calico--apiserver--57bc98fcf9--qfvhc-eth0" Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.174 [INFO][6352] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 7 01:50:31.177960 containerd[1817]: 2026-03-07 01:50:31.176 [INFO][6321] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6" Mar 7 01:50:31.178390 containerd[1817]: time="2026-03-07T01:50:31.178025206Z" level=info msg="TearDown network for sandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\" successfully" Mar 7 01:50:31.186691 containerd[1817]: time="2026-03-07T01:50:31.186644672Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 7 01:50:31.186818 containerd[1817]: time="2026-03-07T01:50:31.186720352Z" level=info msg="RemovePodSandbox \"6c698318b3c49c5d1c6b4753617b9c3027517fd4f175fd3e54aab20ba6b97cc6\" returns successfully" Mar 7 01:50:31.366693 containerd[1817]: time="2026-03-07T01:50:31.366639923Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 01:50:31.371418 containerd[1817]: time="2026-03-07T01:50:31.370327893Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=77" Mar 7 01:50:31.373899 containerd[1817]: time="2026-03-07T01:50:31.373852504Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:dca640051f09574f3e8821035bbfae8c638fb7dadca4c9a082e7223a234befc8\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"46949856\" in 360.116583ms" Mar 7 01:50:31.373899 containerd[1817]: time="2026-03-07T01:50:31.373899944Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:dca640051f09574f3e8821035bbfae8c638fb7dadca4c9a082e7223a234befc8\"" Mar 7 01:50:31.382229 containerd[1817]: time="2026-03-07T01:50:31.382191848Z" level=info msg="CreateContainer within sandbox \"dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 7 01:50:31.421378 containerd[1817]: time="2026-03-07T01:50:31.421334964Z" level=info msg="CreateContainer within sandbox \"dd677c3c660b721af99a35903448d14f9e7867d6b41be5e816b06eb240166b01\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"818b47dfa1bb966c0d32b24e75a8fffa6304fa48987a7d9e43f853b479c8ee26\"" Mar 7 01:50:31.422259 containerd[1817]: time="2026-03-07T01:50:31.422229487Z" level=info msg="StartContainer for \"818b47dfa1bb966c0d32b24e75a8fffa6304fa48987a7d9e43f853b479c8ee26\"" Mar 7 01:50:31.527505 containerd[1817]: time="2026-03-07T01:50:31.527415517Z" level=info msg="StartContainer for \"818b47dfa1bb966c0d32b24e75a8fffa6304fa48987a7d9e43f853b479c8ee26\" returns successfully" Mar 7 01:50:31.845483 kubelet[3332]: I0307 01:50:31.843542 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-68c448d44-wz8jw" podStartSLOduration=36.357140563 podStartE2EDuration="44.84352269s" podCreationTimestamp="2026-03-07 01:49:47 +0000 UTC" firstStartedPulling="2026-03-07 01:50:22.525119788 +0000 UTC m=+54.072721824" lastFinishedPulling="2026-03-07 01:50:31.011501915 +0000 UTC m=+62.559103951" observedRunningTime="2026-03-07 01:50:31.841112726 +0000 UTC m=+63.388714762" watchObservedRunningTime="2026-03-07 01:50:31.84352269 +0000 UTC m=+63.391124686" Mar 7 01:50:31.902685 kubelet[3332]: I0307 01:50:31.901414 3332 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-57bc98fcf9-qfvhc" podStartSLOduration=38.148771954 podStartE2EDuration="46.901395236s" podCreationTimestamp="2026-03-07 01:49:45 +0000 UTC" firstStartedPulling="2026-03-07 01:50:22.621993984 +0000 UTC m=+54.169596020" lastFinishedPulling="2026-03-07 01:50:31.374617306 +0000 UTC m=+62.922219302" observedRunningTime="2026-03-07 01:50:31.860946562 +0000 UTC m=+63.408548598" watchObservedRunningTime="2026-03-07 01:50:31.901395236 +0000 UTC m=+63.448997232" Mar 7 01:50:32.022301 systemd[1]: run-containerd-runc-k8s.io-471aa9f8cf9f70e5a5957091dbe6d6abe46a2ff89874d943b91ed33c0dc31e39-runc.jDVeIQ.mount: Deactivated successfully. Mar 7 01:50:32.835025 kubelet[3332]: I0307 01:50:32.834980 3332 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:50:41.420312 kubelet[3332]: I0307 01:50:41.420278 3332 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:50:57.475696 kubelet[3332]: I0307 01:50:57.475610 3332 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 01:51:13.651651 systemd[1]: Started sshd@7-10.200.20.15:22-10.200.16.10:49570.service - OpenSSH per-connection server daemon (10.200.16.10:49570). Mar 7 01:51:14.155769 sshd[6610]: Accepted publickey for core from 10.200.16.10 port 49570 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:14.157633 sshd[6610]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:14.162407 systemd-logind[1784]: New session 10 of user core. Mar 7 01:51:14.163736 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 7 01:51:14.610388 sshd[6610]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:14.615868 systemd[1]: sshd@7-10.200.20.15:22-10.200.16.10:49570.service: Deactivated successfully. Mar 7 01:51:14.619098 systemd-logind[1784]: Session 10 logged out. Waiting for processes to exit. Mar 7 01:51:14.619862 systemd[1]: session-10.scope: Deactivated successfully. Mar 7 01:51:14.621363 systemd-logind[1784]: Removed session 10. Mar 7 01:51:19.694674 systemd[1]: Started sshd@8-10.200.20.15:22-10.200.16.10:49578.service - OpenSSH per-connection server daemon (10.200.16.10:49578). Mar 7 01:51:20.181206 sshd[6645]: Accepted publickey for core from 10.200.16.10 port 49578 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:20.182569 sshd[6645]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:20.186969 systemd-logind[1784]: New session 11 of user core. Mar 7 01:51:20.191668 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 7 01:51:20.597670 sshd[6645]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:20.602045 systemd[1]: sshd@8-10.200.20.15:22-10.200.16.10:49578.service: Deactivated successfully. Mar 7 01:51:20.607232 systemd[1]: session-11.scope: Deactivated successfully. Mar 7 01:51:20.607950 systemd-logind[1784]: Session 11 logged out. Waiting for processes to exit. Mar 7 01:51:20.608805 systemd-logind[1784]: Removed session 11. Mar 7 01:51:25.682695 systemd[1]: Started sshd@9-10.200.20.15:22-10.200.16.10:44812.service - OpenSSH per-connection server daemon (10.200.16.10:44812). Mar 7 01:51:26.173077 sshd[6660]: Accepted publickey for core from 10.200.16.10 port 44812 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:26.173944 sshd[6660]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:26.178079 systemd-logind[1784]: New session 12 of user core. Mar 7 01:51:26.187694 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 7 01:51:26.585309 sshd[6660]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:26.588207 systemd[1]: sshd@9-10.200.20.15:22-10.200.16.10:44812.service: Deactivated successfully. Mar 7 01:51:26.591692 systemd[1]: session-12.scope: Deactivated successfully. Mar 7 01:51:26.593317 systemd-logind[1784]: Session 12 logged out. Waiting for processes to exit. Mar 7 01:51:26.594600 systemd-logind[1784]: Removed session 12. Mar 7 01:51:31.671670 systemd[1]: Started sshd@10-10.200.20.15:22-10.200.16.10:36396.service - OpenSSH per-connection server daemon (10.200.16.10:36396). Mar 7 01:51:32.159997 sshd[6711]: Accepted publickey for core from 10.200.16.10 port 36396 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:32.161368 sshd[6711]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:32.167721 systemd-logind[1784]: New session 13 of user core. Mar 7 01:51:32.171739 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 7 01:51:32.574436 sshd[6711]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:32.577845 systemd-logind[1784]: Session 13 logged out. Waiting for processes to exit. Mar 7 01:51:32.578082 systemd[1]: sshd@10-10.200.20.15:22-10.200.16.10:36396.service: Deactivated successfully. Mar 7 01:51:32.582405 systemd[1]: session-13.scope: Deactivated successfully. Mar 7 01:51:32.583818 systemd-logind[1784]: Removed session 13. Mar 7 01:51:37.661229 systemd[1]: Started sshd@11-10.200.20.15:22-10.200.16.10:36412.service - OpenSSH per-connection server daemon (10.200.16.10:36412). Mar 7 01:51:38.153264 sshd[6767]: Accepted publickey for core from 10.200.16.10 port 36412 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:38.154254 sshd[6767]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:38.158601 systemd-logind[1784]: New session 14 of user core. Mar 7 01:51:38.162751 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 7 01:51:38.567654 sshd[6767]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:38.570466 systemd[1]: sshd@11-10.200.20.15:22-10.200.16.10:36412.service: Deactivated successfully. Mar 7 01:51:38.574323 systemd-logind[1784]: Session 14 logged out. Waiting for processes to exit. Mar 7 01:51:38.574610 systemd[1]: session-14.scope: Deactivated successfully. Mar 7 01:51:38.576335 systemd-logind[1784]: Removed session 14. Mar 7 01:51:38.658672 systemd[1]: Started sshd@12-10.200.20.15:22-10.200.16.10:36416.service - OpenSSH per-connection server daemon (10.200.16.10:36416). Mar 7 01:51:39.147195 sshd[6782]: Accepted publickey for core from 10.200.16.10 port 36416 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:39.148928 sshd[6782]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:39.152779 systemd-logind[1784]: New session 15 of user core. Mar 7 01:51:39.159715 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 7 01:51:39.599732 sshd[6782]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:39.604765 systemd[1]: sshd@12-10.200.20.15:22-10.200.16.10:36416.service: Deactivated successfully. Mar 7 01:51:39.607714 systemd[1]: session-15.scope: Deactivated successfully. Mar 7 01:51:39.608737 systemd-logind[1784]: Session 15 logged out. Waiting for processes to exit. Mar 7 01:51:39.609659 systemd-logind[1784]: Removed session 15. Mar 7 01:51:39.688705 systemd[1]: Started sshd@13-10.200.20.15:22-10.200.16.10:36420.service - OpenSSH per-connection server daemon (10.200.16.10:36420). Mar 7 01:51:40.173824 sshd[6794]: Accepted publickey for core from 10.200.16.10 port 36420 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:40.175602 sshd[6794]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:40.179875 systemd-logind[1784]: New session 16 of user core. Mar 7 01:51:40.185829 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 7 01:51:40.599689 sshd[6794]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:40.603865 systemd[1]: sshd@13-10.200.20.15:22-10.200.16.10:36420.service: Deactivated successfully. Mar 7 01:51:40.606873 systemd[1]: session-16.scope: Deactivated successfully. Mar 7 01:51:40.608106 systemd-logind[1784]: Session 16 logged out. Waiting for processes to exit. Mar 7 01:51:40.609046 systemd-logind[1784]: Removed session 16. Mar 7 01:51:43.222821 systemd[1]: run-containerd-runc-k8s.io-471aa9f8cf9f70e5a5957091dbe6d6abe46a2ff89874d943b91ed33c0dc31e39-runc.hdyhlb.mount: Deactivated successfully. Mar 7 01:51:45.690753 systemd[1]: Started sshd@14-10.200.20.15:22-10.200.16.10:58192.service - OpenSSH per-connection server daemon (10.200.16.10:58192). Mar 7 01:51:46.180506 sshd[6898]: Accepted publickey for core from 10.200.16.10 port 58192 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:46.208527 sshd[6898]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:46.212204 systemd-logind[1784]: New session 17 of user core. Mar 7 01:51:46.219689 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 7 01:51:46.600699 sshd[6898]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:46.605246 systemd[1]: sshd@14-10.200.20.15:22-10.200.16.10:58192.service: Deactivated successfully. Mar 7 01:51:46.608380 systemd[1]: session-17.scope: Deactivated successfully. Mar 7 01:51:46.609542 systemd-logind[1784]: Session 17 logged out. Waiting for processes to exit. Mar 7 01:51:46.610668 systemd-logind[1784]: Removed session 17. Mar 7 01:51:46.685851 systemd[1]: Started sshd@15-10.200.20.15:22-10.200.16.10:58198.service - OpenSSH per-connection server daemon (10.200.16.10:58198). Mar 7 01:51:47.174080 sshd[6911]: Accepted publickey for core from 10.200.16.10 port 58198 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:47.175515 sshd[6911]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:47.179094 systemd-logind[1784]: New session 18 of user core. Mar 7 01:51:47.185742 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 7 01:51:47.743780 sshd[6911]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:47.747283 systemd-logind[1784]: Session 18 logged out. Waiting for processes to exit. Mar 7 01:51:47.749403 systemd[1]: sshd@15-10.200.20.15:22-10.200.16.10:58198.service: Deactivated successfully. Mar 7 01:51:47.753211 systemd[1]: session-18.scope: Deactivated successfully. Mar 7 01:51:47.755381 systemd-logind[1784]: Removed session 18. Mar 7 01:51:47.829694 systemd[1]: Started sshd@16-10.200.20.15:22-10.200.16.10:58204.service - OpenSSH per-connection server daemon (10.200.16.10:58204). Mar 7 01:51:48.319461 sshd[6935]: Accepted publickey for core from 10.200.16.10 port 58204 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:48.320590 sshd[6935]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:48.324946 systemd-logind[1784]: New session 19 of user core. Mar 7 01:51:48.332720 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 7 01:51:49.508733 sshd[6935]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:49.512447 systemd-logind[1784]: Session 19 logged out. Waiting for processes to exit. Mar 7 01:51:49.513257 systemd[1]: sshd@16-10.200.20.15:22-10.200.16.10:58204.service: Deactivated successfully. Mar 7 01:51:49.517344 systemd[1]: session-19.scope: Deactivated successfully. Mar 7 01:51:49.518695 systemd-logind[1784]: Removed session 19. Mar 7 01:51:49.594656 systemd[1]: Started sshd@17-10.200.20.15:22-10.200.16.10:58216.service - OpenSSH per-connection server daemon (10.200.16.10:58216). Mar 7 01:51:50.080302 sshd[6968]: Accepted publickey for core from 10.200.16.10 port 58216 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:50.081771 sshd[6968]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:50.085634 systemd-logind[1784]: New session 20 of user core. Mar 7 01:51:50.091802 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 7 01:51:50.613668 sshd[6968]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:50.617338 systemd[1]: sshd@17-10.200.20.15:22-10.200.16.10:58216.service: Deactivated successfully. Mar 7 01:51:50.621170 systemd[1]: session-20.scope: Deactivated successfully. Mar 7 01:51:50.622081 systemd-logind[1784]: Session 20 logged out. Waiting for processes to exit. Mar 7 01:51:50.623332 systemd-logind[1784]: Removed session 20. Mar 7 01:51:50.702814 systemd[1]: Started sshd@18-10.200.20.15:22-10.200.16.10:60462.service - OpenSSH per-connection server daemon (10.200.16.10:60462). Mar 7 01:51:51.187745 sshd[6979]: Accepted publickey for core from 10.200.16.10 port 60462 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:51.189048 sshd[6979]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:51.192915 systemd-logind[1784]: New session 21 of user core. Mar 7 01:51:51.199656 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 7 01:51:51.606218 sshd[6979]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:51.610401 systemd[1]: sshd@18-10.200.20.15:22-10.200.16.10:60462.service: Deactivated successfully. Mar 7 01:51:51.614289 systemd[1]: session-21.scope: Deactivated successfully. Mar 7 01:51:51.615165 systemd-logind[1784]: Session 21 logged out. Waiting for processes to exit. Mar 7 01:51:51.616033 systemd-logind[1784]: Removed session 21. Mar 7 01:51:56.691681 systemd[1]: Started sshd@19-10.200.20.15:22-10.200.16.10:60470.service - OpenSSH per-connection server daemon (10.200.16.10:60470). Mar 7 01:51:57.179478 sshd[7014]: Accepted publickey for core from 10.200.16.10 port 60470 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:51:57.180995 sshd[7014]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:51:57.185734 systemd-logind[1784]: New session 22 of user core. Mar 7 01:51:57.192724 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 7 01:51:57.598675 sshd[7014]: pam_unix(sshd:session): session closed for user core Mar 7 01:51:57.601544 systemd-logind[1784]: Session 22 logged out. Waiting for processes to exit. Mar 7 01:51:57.601883 systemd[1]: sshd@19-10.200.20.15:22-10.200.16.10:60470.service: Deactivated successfully. Mar 7 01:51:57.605423 systemd[1]: session-22.scope: Deactivated successfully. Mar 7 01:51:57.607753 systemd-logind[1784]: Removed session 22. Mar 7 01:52:02.684663 systemd[1]: Started sshd@20-10.200.20.15:22-10.200.16.10:46746.service - OpenSSH per-connection server daemon (10.200.16.10:46746). Mar 7 01:52:03.172956 sshd[7048]: Accepted publickey for core from 10.200.16.10 port 46746 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:52:03.174678 sshd[7048]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:52:03.178524 systemd-logind[1784]: New session 23 of user core. Mar 7 01:52:03.182789 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 7 01:52:03.584561 sshd[7048]: pam_unix(sshd:session): session closed for user core Mar 7 01:52:03.588341 systemd[1]: sshd@20-10.200.20.15:22-10.200.16.10:46746.service: Deactivated successfully. Mar 7 01:52:03.591877 systemd[1]: session-23.scope: Deactivated successfully. Mar 7 01:52:03.592098 systemd-logind[1784]: Session 23 logged out. Waiting for processes to exit. Mar 7 01:52:03.593947 systemd-logind[1784]: Removed session 23. Mar 7 01:52:08.669669 systemd[1]: Started sshd@21-10.200.20.15:22-10.200.16.10:46762.service - OpenSSH per-connection server daemon (10.200.16.10:46762). Mar 7 01:52:09.153945 sshd[7064]: Accepted publickey for core from 10.200.16.10 port 46762 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:52:09.155065 sshd[7064]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:52:09.159040 systemd-logind[1784]: New session 24 of user core. Mar 7 01:52:09.164699 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 7 01:52:09.560212 sshd[7064]: pam_unix(sshd:session): session closed for user core Mar 7 01:52:09.564617 systemd-logind[1784]: Session 24 logged out. Waiting for processes to exit. Mar 7 01:52:09.565226 systemd[1]: sshd@21-10.200.20.15:22-10.200.16.10:46762.service: Deactivated successfully. Mar 7 01:52:09.568094 systemd[1]: session-24.scope: Deactivated successfully. Mar 7 01:52:09.569762 systemd-logind[1784]: Removed session 24. Mar 7 01:52:14.647670 systemd[1]: Started sshd@22-10.200.20.15:22-10.200.16.10:43908.service - OpenSSH per-connection server daemon (10.200.16.10:43908). Mar 7 01:52:15.137039 sshd[7079]: Accepted publickey for core from 10.200.16.10 port 43908 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:52:15.138465 sshd[7079]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:52:15.142336 systemd-logind[1784]: New session 25 of user core. Mar 7 01:52:15.147668 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 7 01:52:15.547962 sshd[7079]: pam_unix(sshd:session): session closed for user core Mar 7 01:52:15.552215 systemd-logind[1784]: Session 25 logged out. Waiting for processes to exit. Mar 7 01:52:15.556356 systemd[1]: sshd@22-10.200.20.15:22-10.200.16.10:43908.service: Deactivated successfully. Mar 7 01:52:15.559666 systemd[1]: session-25.scope: Deactivated successfully. Mar 7 01:52:15.560843 systemd-logind[1784]: Removed session 25. Mar 7 01:52:20.653799 systemd[1]: Started sshd@23-10.200.20.15:22-10.200.16.10:38092.service - OpenSSH per-connection server daemon (10.200.16.10:38092). Mar 7 01:52:21.143836 sshd[7114]: Accepted publickey for core from 10.200.16.10 port 38092 ssh2: RSA SHA256:DLehdckcD/qASuOCyInlRssdzQYgoF60IcG9a7M8L7Q Mar 7 01:52:21.145252 sshd[7114]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 01:52:21.149338 systemd-logind[1784]: New session 26 of user core. Mar 7 01:52:21.154682 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 7 01:52:21.566083 sshd[7114]: pam_unix(sshd:session): session closed for user core Mar 7 01:52:21.572532 systemd-logind[1784]: Session 26 logged out. Waiting for processes to exit. Mar 7 01:52:21.573120 systemd[1]: sshd@23-10.200.20.15:22-10.200.16.10:38092.service: Deactivated successfully. Mar 7 01:52:21.575858 systemd[1]: session-26.scope: Deactivated successfully. Mar 7 01:52:21.578436 systemd-logind[1784]: Removed session 26.