Mar 10 01:29:38.036196 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Mar 9 22:55:40 -00 2026 Mar 10 01:29:38.036228 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2de2345ba8612ade61882513e7d9ebf4aad52996b6d7f4c567d9970e886b17cc Mar 10 01:29:38.036250 kernel: BIOS-provided physical RAM map: Mar 10 01:29:38.036264 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Mar 10 01:29:38.036274 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Mar 10 01:29:38.036284 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Mar 10 01:29:38.036295 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdbfff] usable Mar 10 01:29:38.036305 kernel: BIOS-e820: [mem 0x000000007ffdc000-0x000000007fffffff] reserved Mar 10 01:29:38.036315 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Mar 10 01:29:38.036325 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Mar 10 01:29:38.036335 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 10 01:29:38.036345 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Mar 10 01:29:38.036384 kernel: NX (Execute Disable) protection: active Mar 10 01:29:38.036408 kernel: APIC: Static calls initialized Mar 10 01:29:38.036421 kernel: SMBIOS 2.8 present. Mar 10 01:29:38.036432 kernel: DMI: Red Hat KVM/RHEL-AV, BIOS 1.13.0-2.module_el8.5.0+2608+72063365 04/01/2014 Mar 10 01:29:38.036443 kernel: Hypervisor detected: KVM Mar 10 01:29:38.036471 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 10 01:29:38.036483 kernel: kvm-clock: using sched offset of 4474485216 cycles Mar 10 01:29:38.036494 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 10 01:29:38.036506 kernel: tsc: Detected 2799.998 MHz processor Mar 10 01:29:38.036517 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 10 01:29:38.036539 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 10 01:29:38.036557 kernel: last_pfn = 0x7ffdc max_arch_pfn = 0x400000000 Mar 10 01:29:38.036568 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Mar 10 01:29:38.036579 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 10 01:29:38.036595 kernel: Using GB pages for direct mapping Mar 10 01:29:38.036607 kernel: ACPI: Early table checksum verification disabled Mar 10 01:29:38.036618 kernel: ACPI: RSDP 0x00000000000F5AA0 000014 (v00 BOCHS ) Mar 10 01:29:38.036629 kernel: ACPI: RSDT 0x000000007FFE47A5 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 10 01:29:38.036640 kernel: ACPI: FACP 0x000000007FFE438D 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 10 01:29:38.036652 kernel: ACPI: DSDT 0x000000007FFDFD80 00460D (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 10 01:29:38.036662 kernel: ACPI: FACS 0x000000007FFDFD40 000040 Mar 10 01:29:38.036673 kernel: ACPI: APIC 0x000000007FFE4481 0000F0 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 10 01:29:38.036684 kernel: ACPI: SRAT 0x000000007FFE4571 0001D0 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 10 01:29:38.036699 kernel: ACPI: MCFG 0x000000007FFE4741 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 10 01:29:38.036711 kernel: ACPI: WAET 0x000000007FFE477D 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 10 01:29:38.036722 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe438d-0x7ffe4480] Mar 10 01:29:38.036733 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffdfd80-0x7ffe438c] Mar 10 01:29:38.036744 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffdfd40-0x7ffdfd7f] Mar 10 01:29:38.036761 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe4481-0x7ffe4570] Mar 10 01:29:38.036773 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe4571-0x7ffe4740] Mar 10 01:29:38.036788 kernel: ACPI: Reserving MCFG table memory at [mem 0x7ffe4741-0x7ffe477c] Mar 10 01:29:38.036800 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe477d-0x7ffe47a4] Mar 10 01:29:38.036823 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Mar 10 01:29:38.036833 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Mar 10 01:29:38.036844 kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0 Mar 10 01:29:38.036854 kernel: SRAT: PXM 0 -> APIC 0x03 -> Node 0 Mar 10 01:29:38.036864 kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0 Mar 10 01:29:38.036895 kernel: SRAT: PXM 0 -> APIC 0x05 -> Node 0 Mar 10 01:29:38.036910 kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0 Mar 10 01:29:38.036920 kernel: SRAT: PXM 0 -> APIC 0x07 -> Node 0 Mar 10 01:29:38.036930 kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0 Mar 10 01:29:38.036952 kernel: SRAT: PXM 0 -> APIC 0x09 -> Node 0 Mar 10 01:29:38.036963 kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0 Mar 10 01:29:38.036973 kernel: SRAT: PXM 0 -> APIC 0x0b -> Node 0 Mar 10 01:29:38.036984 kernel: SRAT: PXM 0 -> APIC 0x0c -> Node 0 Mar 10 01:29:38.037007 kernel: SRAT: PXM 0 -> APIC 0x0d -> Node 0 Mar 10 01:29:38.037018 kernel: SRAT: PXM 0 -> APIC 0x0e -> Node 0 Mar 10 01:29:38.037033 kernel: SRAT: PXM 0 -> APIC 0x0f -> Node 0 Mar 10 01:29:38.037044 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Mar 10 01:29:38.037055 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Mar 10 01:29:38.037079 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x20800fffff] hotplug Mar 10 01:29:38.037091 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdbfff] -> [mem 0x00000000-0x7ffdbfff] Mar 10 01:29:38.037103 kernel: NODE_DATA(0) allocated [mem 0x7ffd6000-0x7ffdbfff] Mar 10 01:29:38.037114 kernel: Zone ranges: Mar 10 01:29:38.037126 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 10 01:29:38.037137 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdbfff] Mar 10 01:29:38.037153 kernel: Normal empty Mar 10 01:29:38.037164 kernel: Movable zone start for each node Mar 10 01:29:38.037176 kernel: Early memory node ranges Mar 10 01:29:38.037187 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Mar 10 01:29:38.037199 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdbfff] Mar 10 01:29:38.037211 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdbfff] Mar 10 01:29:38.037222 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 10 01:29:38.037233 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Mar 10 01:29:38.037245 kernel: On node 0, zone DMA32: 36 pages in unavailable ranges Mar 10 01:29:38.037256 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 10 01:29:38.037273 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 10 01:29:38.037284 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 10 01:29:38.037296 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 10 01:29:38.037307 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 10 01:29:38.037319 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 10 01:29:38.037330 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 10 01:29:38.037342 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 10 01:29:38.037353 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 10 01:29:38.037369 kernel: TSC deadline timer available Mar 10 01:29:38.037432 kernel: smpboot: Allowing 16 CPUs, 14 hotplug CPUs Mar 10 01:29:38.037446 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 10 01:29:38.037471 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Mar 10 01:29:38.037482 kernel: Booting paravirtualized kernel on KVM Mar 10 01:29:38.037494 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 10 01:29:38.037506 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Mar 10 01:29:38.037518 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u262144 Mar 10 01:29:38.037540 kernel: pcpu-alloc: s196328 r8192 d28952 u262144 alloc=1*2097152 Mar 10 01:29:38.037557 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Mar 10 01:29:38.037575 kernel: kvm-guest: PV spinlocks enabled Mar 10 01:29:38.037587 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 10 01:29:38.037600 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2de2345ba8612ade61882513e7d9ebf4aad52996b6d7f4c567d9970e886b17cc Mar 10 01:29:38.037612 kernel: random: crng init done Mar 10 01:29:38.037623 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 10 01:29:38.037635 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Mar 10 01:29:38.037646 kernel: Fallback order for Node 0: 0 Mar 10 01:29:38.037658 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515804 Mar 10 01:29:38.037674 kernel: Policy zone: DMA32 Mar 10 01:29:38.037686 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 10 01:29:38.037697 kernel: software IO TLB: area num 16. Mar 10 01:29:38.037709 kernel: Memory: 1901592K/2096616K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42896K init, 2300K bss, 194764K reserved, 0K cma-reserved) Mar 10 01:29:38.037720 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Mar 10 01:29:38.037732 kernel: Kernel/User page tables isolation: enabled Mar 10 01:29:38.037743 kernel: ftrace: allocating 37996 entries in 149 pages Mar 10 01:29:38.037755 kernel: ftrace: allocated 149 pages with 4 groups Mar 10 01:29:38.037766 kernel: Dynamic Preempt: voluntary Mar 10 01:29:38.037782 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 10 01:29:38.037795 kernel: rcu: RCU event tracing is enabled. Mar 10 01:29:38.037807 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Mar 10 01:29:38.037818 kernel: Trampoline variant of Tasks RCU enabled. Mar 10 01:29:38.037842 kernel: Rude variant of Tasks RCU enabled. Mar 10 01:29:38.037866 kernel: Tracing variant of Tasks RCU enabled. Mar 10 01:29:38.037882 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 10 01:29:38.037893 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Mar 10 01:29:38.037917 kernel: NR_IRQS: 33024, nr_irqs: 552, preallocated irqs: 16 Mar 10 01:29:38.037929 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 10 01:29:38.037940 kernel: Console: colour VGA+ 80x25 Mar 10 01:29:38.037952 kernel: printk: console [tty0] enabled Mar 10 01:29:38.037981 kernel: printk: console [ttyS0] enabled Mar 10 01:29:38.037993 kernel: ACPI: Core revision 20230628 Mar 10 01:29:38.038005 kernel: APIC: Switch to symmetric I/O mode setup Mar 10 01:29:38.038016 kernel: x2apic enabled Mar 10 01:29:38.038041 kernel: APIC: Switched APIC routing to: physical x2apic Mar 10 01:29:38.038058 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x285c3ee517e, max_idle_ns: 440795257231 ns Mar 10 01:29:38.038070 kernel: Calibrating delay loop (skipped) preset value.. 5599.99 BogoMIPS (lpj=2799998) Mar 10 01:29:38.038082 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 10 01:29:38.038094 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Mar 10 01:29:38.038106 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Mar 10 01:29:38.038118 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 10 01:29:38.038130 kernel: Spectre V2 : Mitigation: Retpolines Mar 10 01:29:38.038148 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Mar 10 01:29:38.038160 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Mar 10 01:29:38.038172 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Mar 10 01:29:38.038188 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Mar 10 01:29:38.038200 kernel: MDS: Mitigation: Clear CPU buffers Mar 10 01:29:38.038212 kernel: MMIO Stale Data: Unknown: No mitigations Mar 10 01:29:38.038224 kernel: SRBDS: Unknown: Dependent on hypervisor status Mar 10 01:29:38.038235 kernel: active return thunk: its_return_thunk Mar 10 01:29:38.038247 kernel: ITS: Mitigation: Aligned branch/return thunks Mar 10 01:29:38.038259 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 10 01:29:38.038271 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 10 01:29:38.038283 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 10 01:29:38.038295 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 10 01:29:38.038307 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Mar 10 01:29:38.038323 kernel: Freeing SMP alternatives memory: 32K Mar 10 01:29:38.038345 kernel: pid_max: default: 32768 minimum: 301 Mar 10 01:29:38.038366 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 10 01:29:38.038399 kernel: landlock: Up and running. Mar 10 01:29:38.038412 kernel: SELinux: Initializing. Mar 10 01:29:38.038424 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Mar 10 01:29:38.038436 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Mar 10 01:29:38.038458 kernel: smpboot: CPU0: Intel Xeon E3-12xx v2 (Ivy Bridge, IBRS) (family: 0x6, model: 0x3a, stepping: 0x9) Mar 10 01:29:38.038470 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Mar 10 01:29:38.038482 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Mar 10 01:29:38.038501 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Mar 10 01:29:38.038519 kernel: Performance Events: unsupported p6 CPU model 58 no PMU driver, software events only. Mar 10 01:29:38.038541 kernel: signal: max sigframe size: 1776 Mar 10 01:29:38.038554 kernel: rcu: Hierarchical SRCU implementation. Mar 10 01:29:38.038567 kernel: rcu: Max phase no-delay instances is 400. Mar 10 01:29:38.038579 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 10 01:29:38.038591 kernel: smp: Bringing up secondary CPUs ... Mar 10 01:29:38.038603 kernel: smpboot: x86: Booting SMP configuration: Mar 10 01:29:38.038615 kernel: .... node #0, CPUs: #1 Mar 10 01:29:38.038627 kernel: smpboot: CPU 1 Converting physical 0 to logical die 1 Mar 10 01:29:38.038645 kernel: smp: Brought up 1 node, 2 CPUs Mar 10 01:29:38.038657 kernel: smpboot: Max logical packages: 16 Mar 10 01:29:38.038669 kernel: smpboot: Total of 2 processors activated (11199.99 BogoMIPS) Mar 10 01:29:38.038681 kernel: devtmpfs: initialized Mar 10 01:29:38.038693 kernel: x86/mm: Memory block size: 128MB Mar 10 01:29:38.038705 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 10 01:29:38.038717 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Mar 10 01:29:38.038729 kernel: pinctrl core: initialized pinctrl subsystem Mar 10 01:29:38.038742 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 10 01:29:38.038758 kernel: audit: initializing netlink subsys (disabled) Mar 10 01:29:38.038771 kernel: audit: type=2000 audit(1773106176.905:1): state=initialized audit_enabled=0 res=1 Mar 10 01:29:38.038782 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 10 01:29:38.038795 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 10 01:29:38.038818 kernel: cpuidle: using governor menu Mar 10 01:29:38.038830 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 10 01:29:38.038841 kernel: dca service started, version 1.12.1 Mar 10 01:29:38.038853 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Mar 10 01:29:38.038877 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Mar 10 01:29:38.038894 kernel: PCI: Using configuration type 1 for base access Mar 10 01:29:38.038906 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 10 01:29:38.038918 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 10 01:29:38.038930 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 10 01:29:38.038942 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 10 01:29:38.038954 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 10 01:29:38.038966 kernel: ACPI: Added _OSI(Module Device) Mar 10 01:29:38.038978 kernel: ACPI: Added _OSI(Processor Device) Mar 10 01:29:38.038995 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 10 01:29:38.039008 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 10 01:29:38.039020 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Mar 10 01:29:38.039032 kernel: ACPI: Interpreter enabled Mar 10 01:29:38.039044 kernel: ACPI: PM: (supports S0 S5) Mar 10 01:29:38.039055 kernel: ACPI: Using IOAPIC for interrupt routing Mar 10 01:29:38.039067 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 10 01:29:38.039079 kernel: PCI: Using E820 reservations for host bridge windows Mar 10 01:29:38.039091 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 10 01:29:38.039103 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 10 01:29:38.039426 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 10 01:29:38.039627 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Mar 10 01:29:38.039858 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Mar 10 01:29:38.039890 kernel: PCI host bridge to bus 0000:00 Mar 10 01:29:38.040097 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 10 01:29:38.040264 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 10 01:29:38.040478 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 10 01:29:38.040681 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window] Mar 10 01:29:38.040883 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Mar 10 01:29:38.041041 kernel: pci_bus 0000:00: root bus resource [mem 0x20c0000000-0x28bfffffff window] Mar 10 01:29:38.041205 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 10 01:29:38.041507 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Mar 10 01:29:38.041730 kernel: pci 0000:00:01.0: [1013:00b8] type 00 class 0x030000 Mar 10 01:29:38.041992 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfa000000-0xfbffffff pref] Mar 10 01:29:38.042177 kernel: pci 0000:00:01.0: reg 0x14: [mem 0xfea50000-0xfea50fff] Mar 10 01:29:38.042379 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfea40000-0xfea4ffff pref] Mar 10 01:29:38.042577 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 10 01:29:38.042793 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.042965 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfea51000-0xfea51fff] Mar 10 01:29:38.043214 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.045508 kernel: pci 0000:00:02.1: reg 0x10: [mem 0xfea52000-0xfea52fff] Mar 10 01:29:38.045729 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.045943 kernel: pci 0000:00:02.2: reg 0x10: [mem 0xfea53000-0xfea53fff] Mar 10 01:29:38.046153 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.046356 kernel: pci 0000:00:02.3: reg 0x10: [mem 0xfea54000-0xfea54fff] Mar 10 01:29:38.046597 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.046769 kernel: pci 0000:00:02.4: reg 0x10: [mem 0xfea55000-0xfea55fff] Mar 10 01:29:38.046982 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.047155 kernel: pci 0000:00:02.5: reg 0x10: [mem 0xfea56000-0xfea56fff] Mar 10 01:29:38.049450 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.049649 kernel: pci 0000:00:02.6: reg 0x10: [mem 0xfea57000-0xfea57fff] Mar 10 01:29:38.049869 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Mar 10 01:29:38.051307 kernel: pci 0000:00:02.7: reg 0x10: [mem 0xfea58000-0xfea58fff] Mar 10 01:29:38.051580 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Mar 10 01:29:38.051753 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc0c0-0xc0df] Mar 10 01:29:38.051921 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfea59000-0xfea59fff] Mar 10 01:29:38.052101 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfd000000-0xfd003fff 64bit pref] Mar 10 01:29:38.052274 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfea00000-0xfea3ffff pref] Mar 10 01:29:38.053579 kernel: pci 0000:00:04.0: [1af4:1001] type 00 class 0x010000 Mar 10 01:29:38.053757 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc07f] Mar 10 01:29:38.053936 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfea5a000-0xfea5afff] Mar 10 01:29:38.054132 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfd004000-0xfd007fff 64bit pref] Mar 10 01:29:38.055961 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Mar 10 01:29:38.056149 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 10 01:29:38.056394 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Mar 10 01:29:38.056592 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc0e0-0xc0ff] Mar 10 01:29:38.056760 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfea5b000-0xfea5bfff] Mar 10 01:29:38.056960 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Mar 10 01:29:38.057130 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Mar 10 01:29:38.058384 kernel: pci 0000:01:00.0: [1b36:000e] type 01 class 0x060400 Mar 10 01:29:38.058594 kernel: pci 0000:01:00.0: reg 0x10: [mem 0xfda00000-0xfda000ff 64bit] Mar 10 01:29:38.058780 kernel: pci 0000:00:02.0: PCI bridge to [bus 01-02] Mar 10 01:29:38.058957 kernel: pci 0000:00:02.0: bridge window [mem 0xfd800000-0xfdbfffff] Mar 10 01:29:38.059134 kernel: pci 0000:00:02.0: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Mar 10 01:29:38.060435 kernel: pci_bus 0000:02: extended config space not accessible Mar 10 01:29:38.060685 kernel: pci 0000:02:01.0: [8086:25ab] type 00 class 0x088000 Mar 10 01:29:38.060878 kernel: pci 0000:02:01.0: reg 0x10: [mem 0xfd800000-0xfd80000f] Mar 10 01:29:38.061061 kernel: pci 0000:01:00.0: PCI bridge to [bus 02] Mar 10 01:29:38.061243 kernel: pci 0000:01:00.0: bridge window [mem 0xfd800000-0xfd9fffff] Mar 10 01:29:38.061467 kernel: pci 0000:03:00.0: [1b36:000d] type 00 class 0x0c0330 Mar 10 01:29:38.061677 kernel: pci 0000:03:00.0: reg 0x10: [mem 0xfe800000-0xfe803fff 64bit] Mar 10 01:29:38.061849 kernel: pci 0000:00:02.1: PCI bridge to [bus 03] Mar 10 01:29:38.062013 kernel: pci 0000:00:02.1: bridge window [mem 0xfe800000-0xfe9fffff] Mar 10 01:29:38.062181 kernel: pci 0000:00:02.1: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Mar 10 01:29:38.064175 kernel: pci 0000:04:00.0: [1af4:1044] type 00 class 0x00ff00 Mar 10 01:29:38.065815 kernel: pci 0000:04:00.0: reg 0x20: [mem 0xfca00000-0xfca03fff 64bit pref] Mar 10 01:29:38.065996 kernel: pci 0000:00:02.2: PCI bridge to [bus 04] Mar 10 01:29:38.066194 kernel: pci 0000:00:02.2: bridge window [mem 0xfe600000-0xfe7fffff] Mar 10 01:29:38.066408 kernel: pci 0000:00:02.2: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Mar 10 01:29:38.066605 kernel: pci 0000:00:02.3: PCI bridge to [bus 05] Mar 10 01:29:38.066772 kernel: pci 0000:00:02.3: bridge window [mem 0xfe400000-0xfe5fffff] Mar 10 01:29:38.066943 kernel: pci 0000:00:02.3: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Mar 10 01:29:38.067130 kernel: pci 0000:00:02.4: PCI bridge to [bus 06] Mar 10 01:29:38.067300 kernel: pci 0000:00:02.4: bridge window [mem 0xfe200000-0xfe3fffff] Mar 10 01:29:38.068549 kernel: pci 0000:00:02.4: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Mar 10 01:29:38.068730 kernel: pci 0000:00:02.5: PCI bridge to [bus 07] Mar 10 01:29:38.068896 kernel: pci 0000:00:02.5: bridge window [mem 0xfe000000-0xfe1fffff] Mar 10 01:29:38.069082 kernel: pci 0000:00:02.5: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Mar 10 01:29:38.069273 kernel: pci 0000:00:02.6: PCI bridge to [bus 08] Mar 10 01:29:38.070713 kernel: pci 0000:00:02.6: bridge window [mem 0xfde00000-0xfdffffff] Mar 10 01:29:38.070893 kernel: pci 0000:00:02.6: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Mar 10 01:29:38.071063 kernel: pci 0000:00:02.7: PCI bridge to [bus 09] Mar 10 01:29:38.071248 kernel: pci 0000:00:02.7: bridge window [mem 0xfdc00000-0xfddfffff] Mar 10 01:29:38.072424 kernel: pci 0000:00:02.7: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Mar 10 01:29:38.072445 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 10 01:29:38.072457 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 10 01:29:38.072469 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 10 01:29:38.072481 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 10 01:29:38.072492 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 10 01:29:38.072535 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 10 01:29:38.072549 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 10 01:29:38.072561 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 10 01:29:38.072574 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 10 01:29:38.072586 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 10 01:29:38.072598 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 10 01:29:38.072610 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 10 01:29:38.072622 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 10 01:29:38.072635 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 10 01:29:38.072653 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 10 01:29:38.072666 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 10 01:29:38.072678 kernel: iommu: Default domain type: Translated Mar 10 01:29:38.072691 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 10 01:29:38.072703 kernel: PCI: Using ACPI for IRQ routing Mar 10 01:29:38.072715 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 10 01:29:38.072728 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Mar 10 01:29:38.072740 kernel: e820: reserve RAM buffer [mem 0x7ffdc000-0x7fffffff] Mar 10 01:29:38.072919 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 10 01:29:38.073097 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 10 01:29:38.073282 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 10 01:29:38.073299 kernel: vgaarb: loaded Mar 10 01:29:38.073320 kernel: clocksource: Switched to clocksource kvm-clock Mar 10 01:29:38.073332 kernel: VFS: Disk quotas dquot_6.6.0 Mar 10 01:29:38.073357 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 10 01:29:38.073369 kernel: pnp: PnP ACPI init Mar 10 01:29:38.074671 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Mar 10 01:29:38.074701 kernel: pnp: PnP ACPI: found 5 devices Mar 10 01:29:38.074714 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 10 01:29:38.074726 kernel: NET: Registered PF_INET protocol family Mar 10 01:29:38.074747 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 10 01:29:38.074760 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Mar 10 01:29:38.074772 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 10 01:29:38.074785 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Mar 10 01:29:38.074797 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Mar 10 01:29:38.074815 kernel: TCP: Hash tables configured (established 16384 bind 16384) Mar 10 01:29:38.074827 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Mar 10 01:29:38.074863 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Mar 10 01:29:38.074875 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 10 01:29:38.074887 kernel: NET: Registered PF_XDP protocol family Mar 10 01:29:38.075083 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01-02] add_size 1000 Mar 10 01:29:38.075261 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Mar 10 01:29:38.076391 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Mar 10 01:29:38.076599 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Mar 10 01:29:38.076767 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Mar 10 01:29:38.076937 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Mar 10 01:29:38.077121 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Mar 10 01:29:38.077285 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Mar 10 01:29:38.077469 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Mar 10 01:29:38.077665 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Mar 10 01:29:38.077840 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Mar 10 01:29:38.078002 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Mar 10 01:29:38.078172 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Mar 10 01:29:38.080392 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Mar 10 01:29:38.080620 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Mar 10 01:29:38.080792 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Mar 10 01:29:38.081017 kernel: pci 0000:01:00.0: PCI bridge to [bus 02] Mar 10 01:29:38.081260 kernel: pci 0000:01:00.0: bridge window [mem 0xfd800000-0xfd9fffff] Mar 10 01:29:38.082490 kernel: pci 0000:00:02.0: PCI bridge to [bus 01-02] Mar 10 01:29:38.082685 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Mar 10 01:29:38.082868 kernel: pci 0000:00:02.0: bridge window [mem 0xfd800000-0xfdbfffff] Mar 10 01:29:38.083032 kernel: pci 0000:00:02.0: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Mar 10 01:29:38.083231 kernel: pci 0000:00:02.1: PCI bridge to [bus 03] Mar 10 01:29:38.084476 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Mar 10 01:29:38.084688 kernel: pci 0000:00:02.1: bridge window [mem 0xfe800000-0xfe9fffff] Mar 10 01:29:38.084854 kernel: pci 0000:00:02.1: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Mar 10 01:29:38.085054 kernel: pci 0000:00:02.2: PCI bridge to [bus 04] Mar 10 01:29:38.085226 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Mar 10 01:29:38.087437 kernel: pci 0000:00:02.2: bridge window [mem 0xfe600000-0xfe7fffff] Mar 10 01:29:38.087631 kernel: pci 0000:00:02.2: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Mar 10 01:29:38.087798 kernel: pci 0000:00:02.3: PCI bridge to [bus 05] Mar 10 01:29:38.088012 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Mar 10 01:29:38.088189 kernel: pci 0000:00:02.3: bridge window [mem 0xfe400000-0xfe5fffff] Mar 10 01:29:38.088353 kernel: pci 0000:00:02.3: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Mar 10 01:29:38.088569 kernel: pci 0000:00:02.4: PCI bridge to [bus 06] Mar 10 01:29:38.088737 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Mar 10 01:29:38.088916 kernel: pci 0000:00:02.4: bridge window [mem 0xfe200000-0xfe3fffff] Mar 10 01:29:38.089093 kernel: pci 0000:00:02.4: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Mar 10 01:29:38.089258 kernel: pci 0000:00:02.5: PCI bridge to [bus 07] Mar 10 01:29:38.089439 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Mar 10 01:29:38.089627 kernel: pci 0000:00:02.5: bridge window [mem 0xfe000000-0xfe1fffff] Mar 10 01:29:38.089795 kernel: pci 0000:00:02.5: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Mar 10 01:29:38.089987 kernel: pci 0000:00:02.6: PCI bridge to [bus 08] Mar 10 01:29:38.090167 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Mar 10 01:29:38.090334 kernel: pci 0000:00:02.6: bridge window [mem 0xfde00000-0xfdffffff] Mar 10 01:29:38.092541 kernel: pci 0000:00:02.6: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Mar 10 01:29:38.092711 kernel: pci 0000:00:02.7: PCI bridge to [bus 09] Mar 10 01:29:38.092898 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Mar 10 01:29:38.093064 kernel: pci 0000:00:02.7: bridge window [mem 0xfdc00000-0xfddfffff] Mar 10 01:29:38.093229 kernel: pci 0000:00:02.7: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Mar 10 01:29:38.095423 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 10 01:29:38.095594 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 10 01:29:38.095745 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 10 01:29:38.095895 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window] Mar 10 01:29:38.096054 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Mar 10 01:29:38.096210 kernel: pci_bus 0000:00: resource 9 [mem 0x20c0000000-0x28bfffffff window] Mar 10 01:29:38.097280 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Mar 10 01:29:38.099483 kernel: pci_bus 0000:01: resource 1 [mem 0xfd800000-0xfdbfffff] Mar 10 01:29:38.099660 kernel: pci_bus 0000:01: resource 2 [mem 0xfce00000-0xfcffffff 64bit pref] Mar 10 01:29:38.099835 kernel: pci_bus 0000:02: resource 1 [mem 0xfd800000-0xfd9fffff] Mar 10 01:29:38.100065 kernel: pci_bus 0000:03: resource 0 [io 0x2000-0x2fff] Mar 10 01:29:38.100353 kernel: pci_bus 0000:03: resource 1 [mem 0xfe800000-0xfe9fffff] Mar 10 01:29:38.100570 kernel: pci_bus 0000:03: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref] Mar 10 01:29:38.100813 kernel: pci_bus 0000:04: resource 0 [io 0x3000-0x3fff] Mar 10 01:29:38.100988 kernel: pci_bus 0000:04: resource 1 [mem 0xfe600000-0xfe7fffff] Mar 10 01:29:38.101163 kernel: pci_bus 0000:04: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref] Mar 10 01:29:38.102403 kernel: pci_bus 0000:05: resource 0 [io 0x4000-0x4fff] Mar 10 01:29:38.102603 kernel: pci_bus 0000:05: resource 1 [mem 0xfe400000-0xfe5fffff] Mar 10 01:29:38.102761 kernel: pci_bus 0000:05: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref] Mar 10 01:29:38.102950 kernel: pci_bus 0000:06: resource 0 [io 0x5000-0x5fff] Mar 10 01:29:38.103144 kernel: pci_bus 0000:06: resource 1 [mem 0xfe200000-0xfe3fffff] Mar 10 01:29:38.103323 kernel: pci_bus 0000:06: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref] Mar 10 01:29:38.104570 kernel: pci_bus 0000:07: resource 0 [io 0x6000-0x6fff] Mar 10 01:29:38.104733 kernel: pci_bus 0000:07: resource 1 [mem 0xfe000000-0xfe1fffff] Mar 10 01:29:38.104900 kernel: pci_bus 0000:07: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref] Mar 10 01:29:38.105100 kernel: pci_bus 0000:08: resource 0 [io 0x7000-0x7fff] Mar 10 01:29:38.105289 kernel: pci_bus 0000:08: resource 1 [mem 0xfde00000-0xfdffffff] Mar 10 01:29:38.105490 kernel: pci_bus 0000:08: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref] Mar 10 01:29:38.105695 kernel: pci_bus 0000:09: resource 0 [io 0x8000-0x8fff] Mar 10 01:29:38.105853 kernel: pci_bus 0000:09: resource 1 [mem 0xfdc00000-0xfddfffff] Mar 10 01:29:38.106047 kernel: pci_bus 0000:09: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref] Mar 10 01:29:38.106075 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 10 01:29:38.106099 kernel: PCI: CLS 0 bytes, default 64 Mar 10 01:29:38.106112 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Mar 10 01:29:38.106126 kernel: software IO TLB: mapped [mem 0x0000000079800000-0x000000007d800000] (64MB) Mar 10 01:29:38.106139 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Mar 10 01:29:38.106152 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x285c3ee517e, max_idle_ns: 440795257231 ns Mar 10 01:29:38.106165 kernel: Initialise system trusted keyrings Mar 10 01:29:38.106179 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Mar 10 01:29:38.106192 kernel: Key type asymmetric registered Mar 10 01:29:38.106211 kernel: Asymmetric key parser 'x509' registered Mar 10 01:29:38.106224 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Mar 10 01:29:38.106237 kernel: io scheduler mq-deadline registered Mar 10 01:29:38.106250 kernel: io scheduler kyber registered Mar 10 01:29:38.106263 kernel: io scheduler bfq registered Mar 10 01:29:38.106510 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Mar 10 01:29:38.106700 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Mar 10 01:29:38.106900 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.107079 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Mar 10 01:29:38.107269 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Mar 10 01:29:38.107455 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.107673 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Mar 10 01:29:38.107852 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Mar 10 01:29:38.108042 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.108259 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Mar 10 01:29:38.108468 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Mar 10 01:29:38.108667 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.108840 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Mar 10 01:29:38.109024 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Mar 10 01:29:38.109270 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.109494 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Mar 10 01:29:38.109695 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Mar 10 01:29:38.109907 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.110125 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Mar 10 01:29:38.110311 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Mar 10 01:29:38.110605 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.110781 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Mar 10 01:29:38.110955 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Mar 10 01:29:38.111165 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 10 01:29:38.111185 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 10 01:29:38.111199 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 10 01:29:38.111211 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Mar 10 01:29:38.111238 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 10 01:29:38.111257 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 10 01:29:38.111270 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 10 01:29:38.111282 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 10 01:29:38.111295 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 10 01:29:38.111308 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 10 01:29:38.111560 kernel: rtc_cmos 00:03: RTC can wake from S4 Mar 10 01:29:38.111723 kernel: rtc_cmos 00:03: registered as rtc0 Mar 10 01:29:38.111911 kernel: rtc_cmos 00:03: setting system clock to 2026-03-10T01:29:37 UTC (1773106177) Mar 10 01:29:38.112077 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Mar 10 01:29:38.112096 kernel: intel_pstate: CPU model not supported Mar 10 01:29:38.112110 kernel: NET: Registered PF_INET6 protocol family Mar 10 01:29:38.112122 kernel: Segment Routing with IPv6 Mar 10 01:29:38.112135 kernel: In-situ OAM (IOAM) with IPv6 Mar 10 01:29:38.112149 kernel: NET: Registered PF_PACKET protocol family Mar 10 01:29:38.112162 kernel: Key type dns_resolver registered Mar 10 01:29:38.112174 kernel: IPI shorthand broadcast: enabled Mar 10 01:29:38.112188 kernel: sched_clock: Marking stable (1266063226, 226818023)->(1622113504, -129232255) Mar 10 01:29:38.112218 kernel: registered taskstats version 1 Mar 10 01:29:38.112232 kernel: Loading compiled-in X.509 certificates Mar 10 01:29:38.112245 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: 611e035accba842cc9fafb5ced2ca41a603067aa' Mar 10 01:29:38.112258 kernel: Key type .fscrypt registered Mar 10 01:29:38.112280 kernel: Key type fscrypt-provisioning registered Mar 10 01:29:38.112293 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 10 01:29:38.112306 kernel: ima: Allocated hash algorithm: sha1 Mar 10 01:29:38.112388 kernel: ima: No architecture policies found Mar 10 01:29:38.112403 kernel: clk: Disabling unused clocks Mar 10 01:29:38.112424 kernel: Freeing unused kernel image (initmem) memory: 42896K Mar 10 01:29:38.112437 kernel: Write protecting the kernel read-only data: 36864k Mar 10 01:29:38.112450 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Mar 10 01:29:38.112463 kernel: Run /init as init process Mar 10 01:29:38.112476 kernel: with arguments: Mar 10 01:29:38.112489 kernel: /init Mar 10 01:29:38.112502 kernel: with environment: Mar 10 01:29:38.112515 kernel: HOME=/ Mar 10 01:29:38.112538 kernel: TERM=linux Mar 10 01:29:38.112561 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 10 01:29:38.112578 systemd[1]: Detected virtualization kvm. Mar 10 01:29:38.112592 systemd[1]: Detected architecture x86-64. Mar 10 01:29:38.112605 systemd[1]: Running in initrd. Mar 10 01:29:38.112618 systemd[1]: No hostname configured, using default hostname. Mar 10 01:29:38.112632 systemd[1]: Hostname set to . Mar 10 01:29:38.112646 systemd[1]: Initializing machine ID from VM UUID. Mar 10 01:29:38.112665 systemd[1]: Queued start job for default target initrd.target. Mar 10 01:29:38.112679 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 10 01:29:38.112693 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 10 01:29:38.112708 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 10 01:29:38.112722 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 10 01:29:38.112736 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 10 01:29:38.112750 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 10 01:29:38.112771 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 10 01:29:38.112786 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 10 01:29:38.112799 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 10 01:29:38.112820 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 10 01:29:38.112833 systemd[1]: Reached target paths.target - Path Units. Mar 10 01:29:38.112847 systemd[1]: Reached target slices.target - Slice Units. Mar 10 01:29:38.112861 systemd[1]: Reached target swap.target - Swaps. Mar 10 01:29:38.112884 systemd[1]: Reached target timers.target - Timer Units. Mar 10 01:29:38.112902 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 10 01:29:38.112921 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 10 01:29:38.112943 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 10 01:29:38.112957 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 10 01:29:38.112971 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 10 01:29:38.112984 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 10 01:29:38.112998 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 10 01:29:38.113012 systemd[1]: Reached target sockets.target - Socket Units. Mar 10 01:29:38.113026 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 10 01:29:38.113045 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 10 01:29:38.113059 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 10 01:29:38.113085 systemd[1]: Starting systemd-fsck-usr.service... Mar 10 01:29:38.113098 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 10 01:29:38.113111 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 10 01:29:38.113187 systemd-journald[202]: Collecting audit messages is disabled. Mar 10 01:29:38.113224 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 10 01:29:38.113246 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 10 01:29:38.113259 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 10 01:29:38.113284 systemd[1]: Finished systemd-fsck-usr.service. Mar 10 01:29:38.113303 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 10 01:29:38.113316 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 10 01:29:38.113328 kernel: Bridge firewalling registered Mar 10 01:29:38.113385 systemd-journald[202]: Journal started Mar 10 01:29:38.113462 systemd-journald[202]: Runtime Journal (/run/log/journal/9e17d4be37ee4163a2d2e321f57b8cd5) is 4.7M, max 38.0M, 33.2M free. Mar 10 01:29:38.068008 systemd-modules-load[203]: Inserted module 'overlay' Mar 10 01:29:38.163842 systemd[1]: Started systemd-journald.service - Journal Service. Mar 10 01:29:38.103359 systemd-modules-load[203]: Inserted module 'br_netfilter' Mar 10 01:29:38.164760 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 10 01:29:38.167836 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 10 01:29:38.178593 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 10 01:29:38.181551 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 10 01:29:38.195132 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 10 01:29:38.196318 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 10 01:29:38.206531 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 10 01:29:38.215450 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 10 01:29:38.217568 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 10 01:29:38.228650 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 10 01:29:38.229744 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 10 01:29:38.244610 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 10 01:29:38.255551 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 10 01:29:38.263220 dracut-cmdline[233]: dracut-dracut-053 Mar 10 01:29:38.268852 dracut-cmdline[233]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2de2345ba8612ade61882513e7d9ebf4aad52996b6d7f4c567d9970e886b17cc Mar 10 01:29:38.305072 systemd-resolved[239]: Positive Trust Anchors: Mar 10 01:29:38.305089 systemd-resolved[239]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 10 01:29:38.305129 systemd-resolved[239]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 10 01:29:38.314477 systemd-resolved[239]: Defaulting to hostname 'linux'. Mar 10 01:29:38.316381 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 10 01:29:38.317539 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 10 01:29:38.372377 kernel: SCSI subsystem initialized Mar 10 01:29:38.384352 kernel: Loading iSCSI transport class v2.0-870. Mar 10 01:29:38.397356 kernel: iscsi: registered transport (tcp) Mar 10 01:29:38.423368 kernel: iscsi: registered transport (qla4xxx) Mar 10 01:29:38.423427 kernel: QLogic iSCSI HBA Driver Mar 10 01:29:38.479322 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 10 01:29:38.487560 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 10 01:29:38.520368 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 10 01:29:38.522653 kernel: device-mapper: uevent: version 1.0.3 Mar 10 01:29:38.522704 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 10 01:29:38.573384 kernel: raid6: sse2x4 gen() 7372 MB/s Mar 10 01:29:38.589410 kernel: raid6: sse2x2 gen() 4841 MB/s Mar 10 01:29:38.608017 kernel: raid6: sse2x1 gen() 4895 MB/s Mar 10 01:29:38.608064 kernel: raid6: using algorithm sse2x4 gen() 7372 MB/s Mar 10 01:29:38.627001 kernel: raid6: .... xor() 4721 MB/s, rmw enabled Mar 10 01:29:38.627054 kernel: raid6: using ssse3x2 recovery algorithm Mar 10 01:29:38.653458 kernel: xor: automatically using best checksumming function avx Mar 10 01:29:38.845399 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 10 01:29:38.860297 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 10 01:29:38.867638 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 10 01:29:38.893522 systemd-udevd[420]: Using default interface naming scheme 'v255'. Mar 10 01:29:38.900730 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 10 01:29:38.909558 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 10 01:29:38.936131 dracut-pre-trigger[426]: rd.md=0: removing MD RAID activation Mar 10 01:29:38.976579 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 10 01:29:38.982554 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 10 01:29:39.115430 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 10 01:29:39.122580 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 10 01:29:39.148663 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 10 01:29:39.155141 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 10 01:29:39.156169 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 10 01:29:39.159534 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 10 01:29:39.167643 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 10 01:29:39.212386 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 10 01:29:39.263278 kernel: virtio_blk virtio1: 2/0/0 default/read/poll queues Mar 10 01:29:39.271351 kernel: cryptd: max_cpu_qlen set to 1000 Mar 10 01:29:39.275358 kernel: virtio_blk virtio1: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Mar 10 01:29:39.304973 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 10 01:29:39.305035 kernel: GPT:17805311 != 125829119 Mar 10 01:29:39.308896 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 10 01:29:39.308942 kernel: GPT:17805311 != 125829119 Mar 10 01:29:39.308972 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 10 01:29:39.310767 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 10 01:29:39.311509 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 10 01:29:39.320558 kernel: AVX version of gcm_enc/dec engaged. Mar 10 01:29:39.320584 kernel: AES CTR mode by8 optimization enabled Mar 10 01:29:39.311724 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 10 01:29:39.317508 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 10 01:29:39.321250 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 10 01:29:39.321457 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 10 01:29:39.339703 kernel: ACPI: bus type USB registered Mar 10 01:29:39.339732 kernel: usbcore: registered new interface driver usbfs Mar 10 01:29:39.339751 kernel: usbcore: registered new interface driver hub Mar 10 01:29:39.339780 kernel: usbcore: registered new device driver usb Mar 10 01:29:39.322916 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 10 01:29:39.332595 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 10 01:29:39.382604 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 10 01:29:39.399505 kernel: libata version 3.00 loaded. Mar 10 01:29:39.415425 kernel: ahci 0000:00:1f.2: version 3.0 Mar 10 01:29:39.415746 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 10 01:29:39.419365 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Mar 10 01:29:39.419640 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 10 01:29:39.423380 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (469) Mar 10 01:29:39.424394 kernel: BTRFS: device fsid a7ce059b-f34b-4785-93b9-44632d452486 devid 1 transid 33 /dev/vda3 scanned by (udev-worker) (483) Mar 10 01:29:39.436178 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 10 01:29:39.537945 kernel: scsi host0: ahci Mar 10 01:29:39.539458 kernel: xhci_hcd 0000:03:00.0: xHCI Host Controller Mar 10 01:29:39.539732 kernel: xhci_hcd 0000:03:00.0: new USB bus registered, assigned bus number 1 Mar 10 01:29:39.539958 kernel: scsi host1: ahci Mar 10 01:29:39.540184 kernel: xhci_hcd 0000:03:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Mar 10 01:29:39.540449 kernel: scsi host2: ahci Mar 10 01:29:39.540681 kernel: xhci_hcd 0000:03:00.0: xHCI Host Controller Mar 10 01:29:39.540892 kernel: xhci_hcd 0000:03:00.0: new USB bus registered, assigned bus number 2 Mar 10 01:29:39.541116 kernel: xhci_hcd 0000:03:00.0: Host supports USB 3.0 SuperSpeed Mar 10 01:29:39.541313 kernel: scsi host3: ahci Mar 10 01:29:39.542216 kernel: scsi host4: ahci Mar 10 01:29:39.542520 kernel: scsi host5: ahci Mar 10 01:29:39.542720 kernel: ata1: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b100 irq 38 Mar 10 01:29:39.542740 kernel: ata2: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b180 irq 38 Mar 10 01:29:39.542757 kernel: ata3: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b200 irq 38 Mar 10 01:29:39.542774 kernel: ata4: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b280 irq 38 Mar 10 01:29:39.542800 kernel: ata5: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b300 irq 38 Mar 10 01:29:39.542818 kernel: ata6: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b380 irq 38 Mar 10 01:29:39.542835 kernel: hub 1-0:1.0: USB hub found Mar 10 01:29:39.543083 kernel: hub 1-0:1.0: 4 ports detected Mar 10 01:29:39.543282 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Mar 10 01:29:39.543555 kernel: hub 2-0:1.0: USB hub found Mar 10 01:29:39.543777 kernel: hub 2-0:1.0: 4 ports detected Mar 10 01:29:39.545036 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 10 01:29:39.552248 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 10 01:29:39.553118 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 10 01:29:39.562216 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 10 01:29:39.567599 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 10 01:29:39.571737 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 10 01:29:39.580924 disk-uuid[562]: Primary Header is updated. Mar 10 01:29:39.580924 disk-uuid[562]: Secondary Entries is updated. Mar 10 01:29:39.580924 disk-uuid[562]: Secondary Header is updated. Mar 10 01:29:39.590697 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 10 01:29:39.597378 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 10 01:29:39.604687 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 10 01:29:39.687400 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Mar 10 01:29:39.756790 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 10 01:29:39.756879 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 10 01:29:39.757526 kernel: ata3: SATA link down (SStatus 0 SControl 300) Mar 10 01:29:39.760370 kernel: ata1: SATA link down (SStatus 0 SControl 300) Mar 10 01:29:39.762346 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 10 01:29:39.765395 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 10 01:29:39.833407 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 10 01:29:39.839507 kernel: usbcore: registered new interface driver usbhid Mar 10 01:29:39.839572 kernel: usbhid: USB HID core driver Mar 10 01:29:39.846824 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:03:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input2 Mar 10 01:29:39.846863 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:03:00.0-1/input0 Mar 10 01:29:40.598733 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 10 01:29:40.599306 disk-uuid[563]: The operation has completed successfully. Mar 10 01:29:40.663052 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 10 01:29:40.663265 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 10 01:29:40.685594 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 10 01:29:40.689588 sh[583]: Success Mar 10 01:29:40.706513 kernel: device-mapper: verity: sha256 using implementation "sha256-avx" Mar 10 01:29:40.771875 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 10 01:29:40.780927 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 10 01:29:40.786477 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 10 01:29:40.820362 kernel: BTRFS info (device dm-0): first mount of filesystem a7ce059b-f34b-4785-93b9-44632d452486 Mar 10 01:29:40.820464 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 10 01:29:40.822525 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 10 01:29:40.824856 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 10 01:29:40.826571 kernel: BTRFS info (device dm-0): using free space tree Mar 10 01:29:40.837213 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 10 01:29:40.838681 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 10 01:29:40.844599 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 10 01:29:40.849508 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 10 01:29:40.863540 kernel: BTRFS info (device vda6): first mount of filesystem 3e73d814-00c9-411d-8220-21b9b3666124 Mar 10 01:29:40.863586 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 10 01:29:40.863605 kernel: BTRFS info (device vda6): using free space tree Mar 10 01:29:40.870355 kernel: BTRFS info (device vda6): auto enabling async discard Mar 10 01:29:40.882149 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 10 01:29:40.884955 kernel: BTRFS info (device vda6): last unmount of filesystem 3e73d814-00c9-411d-8220-21b9b3666124 Mar 10 01:29:40.891174 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 10 01:29:40.898544 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 10 01:29:41.029203 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 10 01:29:41.038667 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 10 01:29:41.050386 ignition[671]: Ignition 2.19.0 Mar 10 01:29:41.050418 ignition[671]: Stage: fetch-offline Mar 10 01:29:41.050517 ignition[671]: no configs at "/usr/lib/ignition/base.d" Mar 10 01:29:41.055408 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 10 01:29:41.050537 ignition[671]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Mar 10 01:29:41.050720 ignition[671]: parsed url from cmdline: "" Mar 10 01:29:41.050727 ignition[671]: no config URL provided Mar 10 01:29:41.050737 ignition[671]: reading system config file "/usr/lib/ignition/user.ign" Mar 10 01:29:41.050752 ignition[671]: no config at "/usr/lib/ignition/user.ign" Mar 10 01:29:41.050761 ignition[671]: failed to fetch config: resource requires networking Mar 10 01:29:41.051304 ignition[671]: Ignition finished successfully Mar 10 01:29:41.080380 systemd-networkd[770]: lo: Link UP Mar 10 01:29:41.080394 systemd-networkd[770]: lo: Gained carrier Mar 10 01:29:41.082795 systemd-networkd[770]: Enumeration completed Mar 10 01:29:41.083690 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 10 01:29:41.083757 systemd-networkd[770]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 10 01:29:41.083762 systemd-networkd[770]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 10 01:29:41.085102 systemd-networkd[770]: eth0: Link UP Mar 10 01:29:41.085108 systemd-networkd[770]: eth0: Gained carrier Mar 10 01:29:41.085140 systemd-networkd[770]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 10 01:29:41.085183 systemd[1]: Reached target network.target - Network. Mar 10 01:29:41.093564 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 10 01:29:41.109479 systemd-networkd[770]: eth0: DHCPv4 address 10.230.50.50/30, gateway 10.230.50.49 acquired from 10.230.50.49 Mar 10 01:29:41.120359 ignition[773]: Ignition 2.19.0 Mar 10 01:29:41.120376 ignition[773]: Stage: fetch Mar 10 01:29:41.120663 ignition[773]: no configs at "/usr/lib/ignition/base.d" Mar 10 01:29:41.120683 ignition[773]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Mar 10 01:29:41.120844 ignition[773]: parsed url from cmdline: "" Mar 10 01:29:41.120850 ignition[773]: no config URL provided Mar 10 01:29:41.120872 ignition[773]: reading system config file "/usr/lib/ignition/user.ign" Mar 10 01:29:41.120887 ignition[773]: no config at "/usr/lib/ignition/user.ign" Mar 10 01:29:41.121027 ignition[773]: config drive ("/dev/disk/by-label/config-2") not found. Waiting... Mar 10 01:29:41.121090 ignition[773]: config drive ("/dev/disk/by-label/CONFIG-2") not found. Waiting... Mar 10 01:29:41.121108 ignition[773]: GET http://169.254.169.254/openstack/latest/user_data: attempt #1 Mar 10 01:29:41.154391 ignition[773]: GET result: OK Mar 10 01:29:41.154954 ignition[773]: parsing config with SHA512: e324f2ac7e4f5e54143aea73dcf91c1885c671a9c15e605b9f5c0bf94b967a691b32cb54fa67419f95a2f06c3e9f2d5b7a83ada9a5d1311da595f7ca51605276 Mar 10 01:29:41.160888 unknown[773]: fetched base config from "system" Mar 10 01:29:41.160904 unknown[773]: fetched base config from "system" Mar 10 01:29:41.161395 ignition[773]: fetch: fetch complete Mar 10 01:29:41.160913 unknown[773]: fetched user config from "openstack" Mar 10 01:29:41.161402 ignition[773]: fetch: fetch passed Mar 10 01:29:41.164574 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 10 01:29:41.161489 ignition[773]: Ignition finished successfully Mar 10 01:29:41.170603 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 10 01:29:41.203349 ignition[780]: Ignition 2.19.0 Mar 10 01:29:41.205304 ignition[780]: Stage: kargs Mar 10 01:29:41.205676 ignition[780]: no configs at "/usr/lib/ignition/base.d" Mar 10 01:29:41.205698 ignition[780]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Mar 10 01:29:41.207068 ignition[780]: kargs: kargs passed Mar 10 01:29:41.207145 ignition[780]: Ignition finished successfully Mar 10 01:29:41.210456 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 10 01:29:41.216541 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 10 01:29:41.238021 ignition[786]: Ignition 2.19.0 Mar 10 01:29:41.238046 ignition[786]: Stage: disks Mar 10 01:29:41.239568 ignition[786]: no configs at "/usr/lib/ignition/base.d" Mar 10 01:29:41.239602 ignition[786]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Mar 10 01:29:41.240970 ignition[786]: disks: disks passed Mar 10 01:29:41.241055 ignition[786]: Ignition finished successfully Mar 10 01:29:41.242342 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 10 01:29:41.243791 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 10 01:29:41.245165 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 10 01:29:41.246824 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 10 01:29:41.248408 systemd[1]: Reached target sysinit.target - System Initialization. Mar 10 01:29:41.249815 systemd[1]: Reached target basic.target - Basic System. Mar 10 01:29:41.258668 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 10 01:29:41.279407 systemd-fsck[794]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Mar 10 01:29:41.282328 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 10 01:29:41.291431 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 10 01:29:41.410389 kernel: EXT4-fs (vda9): mounted filesystem 8ab7565f-94b4-4514-a19e-abd5bcc78da1 r/w with ordered data mode. Quota mode: none. Mar 10 01:29:41.412019 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 10 01:29:41.413658 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 10 01:29:41.421513 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 10 01:29:41.424396 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 10 01:29:41.426518 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 10 01:29:41.431113 systemd[1]: Starting flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent... Mar 10 01:29:41.431959 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 10 01:29:41.432007 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 10 01:29:41.438326 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 10 01:29:41.448482 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (802) Mar 10 01:29:41.448539 kernel: BTRFS info (device vda6): first mount of filesystem 3e73d814-00c9-411d-8220-21b9b3666124 Mar 10 01:29:41.450000 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 10 01:29:41.450046 kernel: BTRFS info (device vda6): using free space tree Mar 10 01:29:41.450936 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 10 01:29:41.460761 kernel: BTRFS info (device vda6): auto enabling async discard Mar 10 01:29:41.474428 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 10 01:29:41.529620 initrd-setup-root[829]: cut: /sysroot/etc/passwd: No such file or directory Mar 10 01:29:41.539234 initrd-setup-root[837]: cut: /sysroot/etc/group: No such file or directory Mar 10 01:29:41.547036 initrd-setup-root[845]: cut: /sysroot/etc/shadow: No such file or directory Mar 10 01:29:41.556277 initrd-setup-root[852]: cut: /sysroot/etc/gshadow: No such file or directory Mar 10 01:29:41.660068 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 10 01:29:41.670497 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 10 01:29:41.674154 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 10 01:29:41.683352 kernel: BTRFS info (device vda6): last unmount of filesystem 3e73d814-00c9-411d-8220-21b9b3666124 Mar 10 01:29:41.715453 ignition[919]: INFO : Ignition 2.19.0 Mar 10 01:29:41.715453 ignition[919]: INFO : Stage: mount Mar 10 01:29:41.719113 ignition[919]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 10 01:29:41.719113 ignition[919]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Mar 10 01:29:41.719113 ignition[919]: INFO : mount: mount passed Mar 10 01:29:41.719113 ignition[919]: INFO : Ignition finished successfully Mar 10 01:29:41.718571 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 10 01:29:41.719692 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 10 01:29:41.819231 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 10 01:29:42.755713 systemd-networkd[770]: eth0: Gained IPv6LL Mar 10 01:29:44.262043 systemd-networkd[770]: eth0: Ignoring DHCPv6 address 2a02:1348:179:8c8c:24:19ff:fee6:3232/128 (valid for 59min 59s, preferred for 59min 59s) which conflicts with 2a02:1348:179:8c8c:24:19ff:fee6:3232/64 assigned by NDisc. Mar 10 01:29:44.262058 systemd-networkd[770]: eth0: Hint: use IPv6Token= setting to change the address generated by NDisc or set UseAutonomousPrefix=no. Mar 10 01:29:48.614050 coreos-metadata[804]: Mar 10 01:29:48.613 WARN failed to locate config-drive, using the metadata service API instead Mar 10 01:29:48.639974 coreos-metadata[804]: Mar 10 01:29:48.639 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Mar 10 01:29:48.653888 coreos-metadata[804]: Mar 10 01:29:48.653 INFO Fetch successful Mar 10 01:29:48.654993 coreos-metadata[804]: Mar 10 01:29:48.654 INFO wrote hostname srv-eb8wl.gb1.brightbox.com to /sysroot/etc/hostname Mar 10 01:29:48.657878 systemd[1]: flatcar-openstack-hostname.service: Deactivated successfully. Mar 10 01:29:48.658104 systemd[1]: Finished flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent. Mar 10 01:29:48.666446 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 10 01:29:48.687532 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 10 01:29:48.699343 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (936) Mar 10 01:29:48.707583 kernel: BTRFS info (device vda6): first mount of filesystem 3e73d814-00c9-411d-8220-21b9b3666124 Mar 10 01:29:48.707630 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 10 01:29:48.707660 kernel: BTRFS info (device vda6): using free space tree Mar 10 01:29:48.712345 kernel: BTRFS info (device vda6): auto enabling async discard Mar 10 01:29:48.715612 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 10 01:29:48.743872 ignition[954]: INFO : Ignition 2.19.0 Mar 10 01:29:48.743872 ignition[954]: INFO : Stage: files Mar 10 01:29:48.745916 ignition[954]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 10 01:29:48.745916 ignition[954]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Mar 10 01:29:48.745916 ignition[954]: DEBUG : files: compiled without relabeling support, skipping Mar 10 01:29:48.749067 ignition[954]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 10 01:29:48.749067 ignition[954]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 10 01:29:48.751645 ignition[954]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 10 01:29:48.751645 ignition[954]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 10 01:29:48.753744 ignition[954]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 10 01:29:48.752563 unknown[954]: wrote ssh authorized keys file for user: core Mar 10 01:29:48.756389 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 10 01:29:48.756389 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Mar 10 01:29:49.045853 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 10 01:29:49.399416 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 10 01:29:49.399416 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 10 01:29:49.399416 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 10 01:29:49.399416 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 10 01:29:49.404656 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.4-x86-64.raw: attempt #1 Mar 10 01:29:49.758189 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Mar 10 01:29:51.450084 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 10 01:29:51.450084 ignition[954]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Mar 10 01:29:51.455343 ignition[954]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 10 01:29:51.455343 ignition[954]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 10 01:29:51.455343 ignition[954]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Mar 10 01:29:51.455343 ignition[954]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Mar 10 01:29:51.455343 ignition[954]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Mar 10 01:29:51.455343 ignition[954]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 10 01:29:51.455343 ignition[954]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 10 01:29:51.455343 ignition[954]: INFO : files: files passed Mar 10 01:29:51.455343 ignition[954]: INFO : Ignition finished successfully Mar 10 01:29:51.455052 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 10 01:29:51.467590 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 10 01:29:51.472479 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 10 01:29:51.477971 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 10 01:29:51.478148 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 10 01:29:51.499824 initrd-setup-root-after-ignition[982]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 10 01:29:51.499824 initrd-setup-root-after-ignition[982]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 10 01:29:51.503483 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 10 01:29:51.506174 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 10 01:29:51.508572 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 10 01:29:51.515517 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 10 01:29:51.551034 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 10 01:29:51.552201 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 10 01:29:51.554587 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 10 01:29:51.555391 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 10 01:29:51.556344 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 10 01:29:51.561553 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 10 01:29:51.581658 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 10 01:29:51.594583 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 10 01:29:51.606818 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 10 01:29:51.607736 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 10 01:29:51.609442 systemd[1]: Stopped target timers.target - Timer Units. Mar 10 01:29:51.610939 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 10 01:29:51.611094 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 10 01:29:51.612981 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 10 01:29:51.614033 systemd[1]: Stopped target basic.target - Basic System. Mar 10 01:29:51.615754 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 10 01:29:51.617126 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 10 01:29:51.618617 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 10 01:29:51.620207 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 10 01:29:51.622003 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 10 01:29:51.623642 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 10 01:29:51.625061 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 10 01:29:51.626694 systemd[1]: Stopped target swap.target - Swaps. Mar 10 01:29:51.627996 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 10 01:29:51.628159 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 10 01:29:51.629978 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 10 01:29:51.630923 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 10 01:29:51.632404 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 10 01:29:51.634702 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 10 01:29:51.635871 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 10 01:29:51.636047 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 10 01:29:51.638059 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 10 01:29:51.638238 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 10 01:29:51.639847 systemd[1]: ignition-files.service: Deactivated successfully. Mar 10 01:29:51.640006 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 10 01:29:51.648539 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 10 01:29:51.651536 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 10 01:29:51.652432 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 10 01:29:51.655259 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 10 01:29:51.656788 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 10 01:29:51.658451 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 10 01:29:51.671010 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 10 01:29:51.671213 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 10 01:29:51.687344 ignition[1006]: INFO : Ignition 2.19.0 Mar 10 01:29:51.687344 ignition[1006]: INFO : Stage: umount Mar 10 01:29:51.687344 ignition[1006]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 10 01:29:51.687344 ignition[1006]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Mar 10 01:29:51.699876 ignition[1006]: INFO : umount: umount passed Mar 10 01:29:51.699876 ignition[1006]: INFO : Ignition finished successfully Mar 10 01:29:51.698532 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 10 01:29:51.698748 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 10 01:29:51.701514 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 10 01:29:51.701603 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 10 01:29:51.702292 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 10 01:29:51.704681 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 10 01:29:51.705739 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 10 01:29:51.705811 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 10 01:29:51.706664 systemd[1]: Stopped target network.target - Network. Mar 10 01:29:51.708446 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 10 01:29:51.708523 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 10 01:29:51.709663 systemd[1]: Stopped target paths.target - Path Units. Mar 10 01:29:51.711138 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 10 01:29:51.715796 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 10 01:29:51.717643 systemd[1]: Stopped target slices.target - Slice Units. Mar 10 01:29:51.718971 systemd[1]: Stopped target sockets.target - Socket Units. Mar 10 01:29:51.720560 systemd[1]: iscsid.socket: Deactivated successfully. Mar 10 01:29:51.720636 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 10 01:29:51.722178 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 10 01:29:51.722254 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 10 01:29:51.723495 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 10 01:29:51.723571 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 10 01:29:51.724854 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 10 01:29:51.724941 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 10 01:29:51.725965 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 10 01:29:51.727552 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 10 01:29:51.730009 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 10 01:29:51.730660 systemd-networkd[770]: eth0: DHCPv6 lease lost Mar 10 01:29:51.733301 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 10 01:29:51.733528 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 10 01:29:51.736506 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 10 01:29:51.736669 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 10 01:29:51.740740 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 10 01:29:51.740822 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 10 01:29:51.748495 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 10 01:29:51.749188 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 10 01:29:51.749266 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 10 01:29:51.750834 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 10 01:29:51.750907 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 10 01:29:51.754083 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 10 01:29:51.754192 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 10 01:29:51.756433 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 10 01:29:51.756521 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 10 01:29:51.759698 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 10 01:29:51.776816 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 10 01:29:51.777849 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 10 01:29:51.781427 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 10 01:29:51.781604 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 10 01:29:51.784002 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 10 01:29:51.784101 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 10 01:29:51.785916 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 10 01:29:51.785968 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 10 01:29:51.787599 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 10 01:29:51.787674 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 10 01:29:51.789737 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 10 01:29:51.789826 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 10 01:29:51.791304 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 10 01:29:51.791410 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 10 01:29:51.799542 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 10 01:29:51.801531 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 10 01:29:51.801606 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 10 01:29:51.803303 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Mar 10 01:29:51.804377 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 10 01:29:51.805166 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 10 01:29:51.805241 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 10 01:29:51.806672 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 10 01:29:51.806745 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 10 01:29:51.811144 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 10 01:29:51.811931 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 10 01:29:51.816560 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 10 01:29:51.816720 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 10 01:29:51.818034 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 10 01:29:51.818862 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 10 01:29:51.818939 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 10 01:29:51.826540 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 10 01:29:51.838887 systemd[1]: Switching root. Mar 10 01:29:51.869799 systemd-journald[202]: Journal stopped Mar 10 01:29:53.378897 systemd-journald[202]: Received SIGTERM from PID 1 (systemd). Mar 10 01:29:53.379087 kernel: SELinux: policy capability network_peer_controls=1 Mar 10 01:29:53.379151 kernel: SELinux: policy capability open_perms=1 Mar 10 01:29:53.379183 kernel: SELinux: policy capability extended_socket_class=1 Mar 10 01:29:53.379207 kernel: SELinux: policy capability always_check_network=0 Mar 10 01:29:53.379245 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 10 01:29:53.379273 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 10 01:29:53.379300 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 10 01:29:53.381353 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 10 01:29:53.381394 systemd[1]: Successfully loaded SELinux policy in 50.981ms. Mar 10 01:29:53.381460 kernel: audit: type=1403 audit(1773106192.120:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 10 01:29:53.381505 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 22.774ms. Mar 10 01:29:53.381527 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 10 01:29:53.381573 systemd[1]: Detected virtualization kvm. Mar 10 01:29:53.381600 systemd[1]: Detected architecture x86-64. Mar 10 01:29:53.381621 systemd[1]: Detected first boot. Mar 10 01:29:53.381646 systemd[1]: Hostname set to . Mar 10 01:29:53.381671 systemd[1]: Initializing machine ID from VM UUID. Mar 10 01:29:53.381710 zram_generator::config[1049]: No configuration found. Mar 10 01:29:53.381731 systemd[1]: Populated /etc with preset unit settings. Mar 10 01:29:53.381759 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 10 01:29:53.381805 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 10 01:29:53.381841 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 10 01:29:53.381880 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 10 01:29:53.381901 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 10 01:29:53.381946 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 10 01:29:53.381972 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 10 01:29:53.381993 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 10 01:29:53.382013 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 10 01:29:53.382039 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 10 01:29:53.382071 systemd[1]: Created slice user.slice - User and Session Slice. Mar 10 01:29:53.382100 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 10 01:29:53.382126 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 10 01:29:53.382171 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 10 01:29:53.382192 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 10 01:29:53.382212 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 10 01:29:53.382233 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 10 01:29:53.382252 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 10 01:29:53.382288 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 10 01:29:53.382310 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 10 01:29:53.383400 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 10 01:29:53.383437 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 10 01:29:53.383461 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 10 01:29:53.383481 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 10 01:29:53.383508 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 10 01:29:53.383553 systemd[1]: Reached target slices.target - Slice Units. Mar 10 01:29:53.383583 systemd[1]: Reached target swap.target - Swaps. Mar 10 01:29:53.383603 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 10 01:29:53.383623 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 10 01:29:53.383652 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 10 01:29:53.383680 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 10 01:29:53.383714 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 10 01:29:53.383747 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 10 01:29:53.383794 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 10 01:29:53.383828 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 10 01:29:53.383849 systemd[1]: Mounting media.mount - External Media Directory... Mar 10 01:29:53.383869 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 10 01:29:53.383889 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 10 01:29:53.383914 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 10 01:29:53.383939 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 10 01:29:53.383971 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 10 01:29:53.384006 systemd[1]: Reached target machines.target - Containers. Mar 10 01:29:53.384032 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 10 01:29:53.384061 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 10 01:29:53.384087 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 10 01:29:53.384113 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 10 01:29:53.384149 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 10 01:29:53.384172 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 10 01:29:53.384214 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 10 01:29:53.384236 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 10 01:29:53.384262 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 10 01:29:53.384283 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 10 01:29:53.384304 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 10 01:29:53.385366 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 10 01:29:53.385400 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 10 01:29:53.385450 systemd[1]: Stopped systemd-fsck-usr.service. Mar 10 01:29:53.385485 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 10 01:29:53.385525 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 10 01:29:53.385545 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 10 01:29:53.385586 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 10 01:29:53.385612 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 10 01:29:53.385650 systemd[1]: verity-setup.service: Deactivated successfully. Mar 10 01:29:53.385671 systemd[1]: Stopped verity-setup.service. Mar 10 01:29:53.385697 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 10 01:29:53.385730 kernel: fuse: init (API version 7.39) Mar 10 01:29:53.385762 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 10 01:29:53.385791 kernel: loop: module loaded Mar 10 01:29:53.385816 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 10 01:29:53.385837 systemd[1]: Mounted media.mount - External Media Directory. Mar 10 01:29:53.385862 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 10 01:29:53.385895 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 10 01:29:53.385917 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 10 01:29:53.385943 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 10 01:29:53.385964 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 10 01:29:53.385984 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 10 01:29:53.386004 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 10 01:29:53.386024 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 10 01:29:53.386056 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 10 01:29:53.386078 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 10 01:29:53.386106 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 10 01:29:53.386125 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 10 01:29:53.386169 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 10 01:29:53.386191 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 10 01:29:53.386210 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 10 01:29:53.386243 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 10 01:29:53.386279 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 10 01:29:53.387432 systemd-journald[1145]: Collecting audit messages is disabled. Mar 10 01:29:53.387520 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 10 01:29:53.387543 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 10 01:29:53.387563 systemd-journald[1145]: Journal started Mar 10 01:29:53.387632 systemd-journald[1145]: Runtime Journal (/run/log/journal/9e17d4be37ee4163a2d2e321f57b8cd5) is 4.7M, max 38.0M, 33.2M free. Mar 10 01:29:52.933933 systemd[1]: Queued start job for default target multi-user.target. Mar 10 01:29:52.957994 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 10 01:29:52.958827 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 10 01:29:53.393360 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 10 01:29:53.410353 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 10 01:29:53.418427 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 10 01:29:53.418504 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 10 01:29:53.422353 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 10 01:29:53.450362 kernel: ACPI: bus type drm_connector registered Mar 10 01:29:53.450471 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 10 01:29:53.459424 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 10 01:29:53.465380 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 10 01:29:53.480348 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 10 01:29:53.489799 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 10 01:29:53.497366 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 10 01:29:53.502353 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 10 01:29:53.516532 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 10 01:29:53.539650 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 10 01:29:53.548582 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 10 01:29:53.555555 systemd[1]: Started systemd-journald.service - Journal Service. Mar 10 01:29:53.558724 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 10 01:29:53.560440 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 10 01:29:53.561591 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 10 01:29:53.563697 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 10 01:29:53.564904 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 10 01:29:53.567027 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 10 01:29:53.615243 kernel: loop0: detected capacity change from 0 to 8 Mar 10 01:29:53.610872 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 10 01:29:53.625534 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 10 01:29:53.638571 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 10 01:29:53.645371 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 10 01:29:53.666901 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 10 01:29:53.694287 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 10 01:29:53.702343 kernel: loop1: detected capacity change from 0 to 140768 Mar 10 01:29:53.697853 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 10 01:29:53.702663 systemd-journald[1145]: Time spent on flushing to /var/log/journal/9e17d4be37ee4163a2d2e321f57b8cd5 is 91.537ms for 1147 entries. Mar 10 01:29:53.702663 systemd-journald[1145]: System Journal (/var/log/journal/9e17d4be37ee4163a2d2e321f57b8cd5) is 8.0M, max 584.8M, 576.8M free. Mar 10 01:29:53.849001 systemd-journald[1145]: Received client request to flush runtime journal. Mar 10 01:29:53.849099 kernel: loop2: detected capacity change from 0 to 142488 Mar 10 01:29:53.707985 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 10 01:29:53.720602 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 10 01:29:53.751908 systemd-tmpfiles[1165]: ACLs are not supported, ignoring. Mar 10 01:29:53.751929 systemd-tmpfiles[1165]: ACLs are not supported, ignoring. Mar 10 01:29:53.790864 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 10 01:29:53.804541 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 10 01:29:53.811576 udevadm[1198]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 10 01:29:53.853532 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 10 01:29:53.902461 kernel: loop3: detected capacity change from 0 to 219192 Mar 10 01:29:53.916884 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 10 01:29:53.935603 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 10 01:29:53.981376 kernel: loop4: detected capacity change from 0 to 8 Mar 10 01:29:53.990206 kernel: loop5: detected capacity change from 0 to 140768 Mar 10 01:29:54.019358 kernel: loop6: detected capacity change from 0 to 142488 Mar 10 01:29:54.032592 systemd-tmpfiles[1207]: ACLs are not supported, ignoring. Mar 10 01:29:54.033492 systemd-tmpfiles[1207]: ACLs are not supported, ignoring. Mar 10 01:29:54.044250 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 10 01:29:54.059566 kernel: loop7: detected capacity change from 0 to 219192 Mar 10 01:29:54.089017 (sd-merge)[1209]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-openstack'. Mar 10 01:29:54.095525 (sd-merge)[1209]: Merged extensions into '/usr'. Mar 10 01:29:54.105276 systemd[1]: Reloading requested from client PID 1164 ('systemd-sysext') (unit systemd-sysext.service)... Mar 10 01:29:54.105831 systemd[1]: Reloading... Mar 10 01:29:54.252358 zram_generator::config[1235]: No configuration found. Mar 10 01:29:54.281368 ldconfig[1160]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 10 01:29:54.462498 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 10 01:29:54.529811 systemd[1]: Reloading finished in 423 ms. Mar 10 01:29:54.562815 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 10 01:29:54.564298 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 10 01:29:54.580615 systemd[1]: Starting ensure-sysext.service... Mar 10 01:29:54.585691 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 10 01:29:54.597921 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 10 01:29:54.607647 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 10 01:29:54.615474 systemd[1]: Reloading requested from client PID 1293 ('systemctl') (unit ensure-sysext.service)... Mar 10 01:29:54.615501 systemd[1]: Reloading... Mar 10 01:29:54.628138 systemd-tmpfiles[1294]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 10 01:29:54.629244 systemd-tmpfiles[1294]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 10 01:29:54.630864 systemd-tmpfiles[1294]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 10 01:29:54.631405 systemd-tmpfiles[1294]: ACLs are not supported, ignoring. Mar 10 01:29:54.631646 systemd-tmpfiles[1294]: ACLs are not supported, ignoring. Mar 10 01:29:54.637287 systemd-tmpfiles[1294]: Detected autofs mount point /boot during canonicalization of boot. Mar 10 01:29:54.637456 systemd-tmpfiles[1294]: Skipping /boot Mar 10 01:29:54.655264 systemd-tmpfiles[1294]: Detected autofs mount point /boot during canonicalization of boot. Mar 10 01:29:54.656208 systemd-tmpfiles[1294]: Skipping /boot Mar 10 01:29:54.696777 systemd-udevd[1296]: Using default interface naming scheme 'v255'. Mar 10 01:29:54.750354 zram_generator::config[1325]: No configuration found. Mar 10 01:29:54.899360 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (1337) Mar 10 01:29:55.044796 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 10 01:29:55.064372 kernel: mousedev: PS/2 mouse device common for all mice Mar 10 01:29:55.118799 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Mar 10 01:29:55.130341 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Mar 10 01:29:55.147349 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 10 01:29:55.155164 kernel: ACPI: button: Power Button [PWRF] Mar 10 01:29:55.155220 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Mar 10 01:29:55.155601 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 10 01:29:55.251892 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 10 01:29:55.254034 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Mar 10 01:29:55.254770 systemd[1]: Reloading finished in 638 ms. Mar 10 01:29:55.291293 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 10 01:29:55.299393 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 10 01:29:55.347730 systemd[1]: Finished ensure-sysext.service. Mar 10 01:29:55.352247 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 10 01:29:55.362505 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 10 01:29:55.395012 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 10 01:29:55.396058 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 10 01:29:55.404549 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 10 01:29:55.408514 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 10 01:29:55.418530 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 10 01:29:55.428037 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 10 01:29:55.430648 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 10 01:29:55.437741 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 10 01:29:55.445525 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 10 01:29:55.465153 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 10 01:29:55.496618 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 10 01:29:55.512780 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 10 01:29:55.521586 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 10 01:29:55.533508 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 10 01:29:55.537429 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 10 01:29:55.539194 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 10 01:29:55.540835 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 10 01:29:55.541593 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 10 01:29:55.543889 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 10 01:29:55.545283 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 10 01:29:55.552444 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 10 01:29:55.556780 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 10 01:29:55.557097 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 10 01:29:55.580536 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 10 01:29:55.581422 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 10 01:29:55.581569 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 10 01:29:55.593537 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 10 01:29:55.594279 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 10 01:29:55.595161 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 10 01:29:55.597101 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 10 01:29:55.598273 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 10 01:29:55.600801 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 10 01:29:55.637968 lvm[1435]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 10 01:29:55.641869 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 10 01:29:55.670396 augenrules[1448]: No rules Mar 10 01:29:55.670631 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 10 01:29:55.673060 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 10 01:29:55.686269 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 10 01:29:55.687460 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 10 01:29:55.696123 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 10 01:29:55.803961 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 10 01:29:55.811002 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 10 01:29:55.812387 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 10 01:29:55.834864 lvm[1457]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 10 01:29:55.851598 systemd-networkd[1419]: lo: Link UP Mar 10 01:29:55.851610 systemd-networkd[1419]: lo: Gained carrier Mar 10 01:29:55.854214 systemd-networkd[1419]: Enumeration completed Mar 10 01:29:55.854399 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 10 01:29:55.857867 systemd-networkd[1419]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 10 01:29:55.857879 systemd-networkd[1419]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 10 01:29:55.864591 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 10 01:29:55.865961 systemd-networkd[1419]: eth0: Link UP Mar 10 01:29:55.865985 systemd-networkd[1419]: eth0: Gained carrier Mar 10 01:29:55.866002 systemd-networkd[1419]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 10 01:29:55.893072 systemd-networkd[1419]: eth0: DHCPv4 address 10.230.50.50/30, gateway 10.230.50.49 acquired from 10.230.50.49 Mar 10 01:29:55.895488 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 10 01:29:55.897641 systemd[1]: Reached target time-set.target - System Time Set. Mar 10 01:29:55.901166 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 10 01:29:55.915919 systemd-resolved[1420]: Positive Trust Anchors: Mar 10 01:29:55.916450 systemd-resolved[1420]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 10 01:29:55.916604 systemd-resolved[1420]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 10 01:29:55.924168 systemd-resolved[1420]: Using system hostname 'srv-eb8wl.gb1.brightbox.com'. Mar 10 01:29:55.927450 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 10 01:29:55.928393 systemd[1]: Reached target network.target - Network. Mar 10 01:29:55.929071 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 10 01:29:55.929918 systemd[1]: Reached target sysinit.target - System Initialization. Mar 10 01:29:55.930781 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 10 01:29:55.931651 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 10 01:29:55.932691 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 10 01:29:55.933615 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 10 01:29:55.934415 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 10 01:29:55.935205 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 10 01:29:55.935253 systemd[1]: Reached target paths.target - Path Units. Mar 10 01:29:55.935950 systemd[1]: Reached target timers.target - Timer Units. Mar 10 01:29:55.937547 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 10 01:29:55.940635 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 10 01:29:55.949134 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 10 01:29:55.950645 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 10 01:29:55.951569 systemd[1]: Reached target sockets.target - Socket Units. Mar 10 01:29:55.952274 systemd[1]: Reached target basic.target - Basic System. Mar 10 01:29:55.953121 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 10 01:29:55.953176 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 10 01:29:55.958469 systemd[1]: Starting containerd.service - containerd container runtime... Mar 10 01:29:55.961582 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 10 01:29:55.968586 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 10 01:29:55.974317 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 10 01:29:55.978002 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 10 01:29:55.978861 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 10 01:29:55.982604 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 10 01:29:55.991466 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 10 01:29:55.997488 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 10 01:29:56.005533 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 10 01:29:56.020909 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 10 01:29:56.023152 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 10 01:29:56.024890 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 10 01:29:56.030504 extend-filesystems[1479]: Found loop4 Mar 10 01:29:56.030504 extend-filesystems[1479]: Found loop5 Mar 10 01:29:56.030504 extend-filesystems[1479]: Found loop6 Mar 10 01:29:56.030504 extend-filesystems[1479]: Found loop7 Mar 10 01:29:56.030504 extend-filesystems[1479]: Found vda Mar 10 01:29:56.030504 extend-filesystems[1479]: Found vda1 Mar 10 01:29:56.030504 extend-filesystems[1479]: Found vda2 Mar 10 01:29:56.047990 extend-filesystems[1479]: Found vda3 Mar 10 01:29:56.047990 extend-filesystems[1479]: Found usr Mar 10 01:29:56.047990 extend-filesystems[1479]: Found vda4 Mar 10 01:29:56.047990 extend-filesystems[1479]: Found vda6 Mar 10 01:29:56.047990 extend-filesystems[1479]: Found vda7 Mar 10 01:29:56.047990 extend-filesystems[1479]: Found vda9 Mar 10 01:29:56.047990 extend-filesystems[1479]: Checking size of /dev/vda9 Mar 10 01:29:56.030535 systemd[1]: Starting update-engine.service - Update Engine... Mar 10 01:29:56.073743 jq[1476]: false Mar 10 01:29:56.042468 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 10 01:29:56.048308 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 10 01:29:56.048647 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 10 01:29:56.083340 jq[1488]: true Mar 10 01:29:56.079865 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 10 01:29:56.100847 update_engine[1486]: I20260310 01:29:56.100701 1486 main.cc:92] Flatcar Update Engine starting Mar 10 01:29:56.113981 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 10 01:29:56.114287 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 10 01:29:56.125950 extend-filesystems[1479]: Resized partition /dev/vda9 Mar 10 01:29:56.137145 jq[1495]: true Mar 10 01:29:56.137500 extend-filesystems[1508]: resize2fs 1.47.1 (20-May-2024) Mar 10 01:29:56.154641 kernel: EXT4-fs (vda9): resizing filesystem from 1617920 to 15121403 blocks Mar 10 01:29:56.162423 (ntainerd)[1507]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 10 01:29:57.074888 tar[1492]: linux-amd64/LICENSE Mar 10 01:29:57.074888 tar[1492]: linux-amd64/helm Mar 10 01:29:56.164504 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 10 01:29:56.163145 dbus-daemon[1475]: [system] SELinux support is enabled Mar 10 01:29:57.074746 systemd-timesyncd[1421]: Contacted time server 85.199.214.99:123 (0.flatcar.pool.ntp.org). Mar 10 01:29:57.074820 systemd-timesyncd[1421]: Initial clock synchronization to Tue 2026-03-10 01:29:57.074538 UTC. Mar 10 01:29:57.081008 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 10 01:29:57.081054 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 10 01:29:57.081668 systemd-resolved[1420]: Clock change detected. Flushing caches. Mar 10 01:29:57.081915 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 10 01:29:57.081949 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 10 01:29:57.094586 dbus-daemon[1475]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1419 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Mar 10 01:29:57.106487 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Mar 10 01:29:57.108841 systemd[1]: Started update-engine.service - Update Engine. Mar 10 01:29:57.117329 update_engine[1486]: I20260310 01:29:57.111744 1486 update_check_scheduler.cc:74] Next update check in 9m44s Mar 10 01:29:57.118417 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 10 01:29:57.155127 systemd[1]: motdgen.service: Deactivated successfully. Mar 10 01:29:57.156114 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 10 01:29:57.254837 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (1328) Mar 10 01:29:57.339287 systemd-logind[1485]: Watching system buttons on /dev/input/event2 (Power Button) Mar 10 01:29:57.339337 systemd-logind[1485]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 10 01:29:57.339997 systemd-logind[1485]: New seat seat0. Mar 10 01:29:57.343332 systemd[1]: Started systemd-logind.service - User Login Management. Mar 10 01:29:57.352247 bash[1534]: Updated "/home/core/.ssh/authorized_keys" Mar 10 01:29:57.355510 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 10 01:29:57.370589 systemd[1]: Starting sshkeys.service... Mar 10 01:29:57.431278 locksmithd[1519]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 10 01:29:57.443118 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Mar 10 01:29:57.457155 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Mar 10 01:29:57.474244 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Mar 10 01:29:57.501136 extend-filesystems[1508]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 10 01:29:57.501136 extend-filesystems[1508]: old_desc_blocks = 1, new_desc_blocks = 8 Mar 10 01:29:57.501136 extend-filesystems[1508]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Mar 10 01:29:57.500833 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 10 01:29:57.515667 extend-filesystems[1479]: Resized filesystem in /dev/vda9 Mar 10 01:29:57.504481 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 10 01:29:57.559969 containerd[1507]: time="2026-03-10T01:29:57.557020929Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Mar 10 01:29:57.577706 dbus-daemon[1475]: [system] Successfully activated service 'org.freedesktop.hostname1' Mar 10 01:29:57.577964 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Mar 10 01:29:57.582506 dbus-daemon[1475]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.8' (uid=0 pid=1517 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Mar 10 01:29:57.595686 systemd[1]: Starting polkit.service - Authorization Manager... Mar 10 01:29:57.630194 containerd[1507]: time="2026-03-10T01:29:57.629920501Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 10 01:29:57.632094 containerd[1507]: time="2026-03-10T01:29:57.632051139Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 10 01:29:57.632094 containerd[1507]: time="2026-03-10T01:29:57.632092078Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 10 01:29:57.632190 containerd[1507]: time="2026-03-10T01:29:57.632115975Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 10 01:29:57.632439 containerd[1507]: time="2026-03-10T01:29:57.632403629Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 10 01:29:57.632511 containerd[1507]: time="2026-03-10T01:29:57.632491673Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 10 01:29:57.632639 containerd[1507]: time="2026-03-10T01:29:57.632614240Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 10 01:29:57.632680 containerd[1507]: time="2026-03-10T01:29:57.632640800Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 10 01:29:57.633242 containerd[1507]: time="2026-03-10T01:29:57.632856982Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 10 01:29:57.633242 containerd[1507]: time="2026-03-10T01:29:57.632886132Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 10 01:29:57.633242 containerd[1507]: time="2026-03-10T01:29:57.632907318Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 10 01:29:57.633242 containerd[1507]: time="2026-03-10T01:29:57.632943704Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 10 01:29:57.633242 containerd[1507]: time="2026-03-10T01:29:57.633084128Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 10 01:29:57.635098 containerd[1507]: time="2026-03-10T01:29:57.633498764Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 10 01:29:57.635098 containerd[1507]: time="2026-03-10T01:29:57.633660385Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 10 01:29:57.635098 containerd[1507]: time="2026-03-10T01:29:57.633691705Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 10 01:29:57.635098 containerd[1507]: time="2026-03-10T01:29:57.633836330Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 10 01:29:57.635098 containerd[1507]: time="2026-03-10T01:29:57.633918861Z" level=info msg="metadata content store policy set" policy=shared Mar 10 01:29:57.634031 polkitd[1552]: Started polkitd version 121 Mar 10 01:29:57.639979 containerd[1507]: time="2026-03-10T01:29:57.639266998Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 10 01:29:57.639979 containerd[1507]: time="2026-03-10T01:29:57.639358944Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 10 01:29:57.639979 containerd[1507]: time="2026-03-10T01:29:57.639397869Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 10 01:29:57.639979 containerd[1507]: time="2026-03-10T01:29:57.639436205Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 10 01:29:57.639979 containerd[1507]: time="2026-03-10T01:29:57.639464580Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 10 01:29:57.639979 containerd[1507]: time="2026-03-10T01:29:57.639656486Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 10 01:29:57.640254 containerd[1507]: time="2026-03-10T01:29:57.640060748Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 10 01:29:57.640328 containerd[1507]: time="2026-03-10T01:29:57.640287492Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 10 01:29:57.640394 containerd[1507]: time="2026-03-10T01:29:57.640333988Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 10 01:29:57.640394 containerd[1507]: time="2026-03-10T01:29:57.640371036Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 10 01:29:57.640466 containerd[1507]: time="2026-03-10T01:29:57.640395060Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640466 containerd[1507]: time="2026-03-10T01:29:57.640417364Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640466 containerd[1507]: time="2026-03-10T01:29:57.640437145Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640466 containerd[1507]: time="2026-03-10T01:29:57.640458862Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640588 containerd[1507]: time="2026-03-10T01:29:57.640480821Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640588 containerd[1507]: time="2026-03-10T01:29:57.640510556Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640588 containerd[1507]: time="2026-03-10T01:29:57.640530551Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640588 containerd[1507]: time="2026-03-10T01:29:57.640548044Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 10 01:29:57.640727 containerd[1507]: time="2026-03-10T01:29:57.640599421Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640727 containerd[1507]: time="2026-03-10T01:29:57.640632056Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640727 containerd[1507]: time="2026-03-10T01:29:57.640664734Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640727 containerd[1507]: time="2026-03-10T01:29:57.640684614Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640727 containerd[1507]: time="2026-03-10T01:29:57.640703064Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640735280Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640755105Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640774297Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640794783Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640817737Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640837420Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640874684Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.640902 containerd[1507]: time="2026-03-10T01:29:57.640892427Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.640938842Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.641003524Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.641026067Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.641047937Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.641113927Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.641147339Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.641166160Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 10 01:29:57.641188 containerd[1507]: time="2026-03-10T01:29:57.641185916Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 10 01:29:57.641452 containerd[1507]: time="2026-03-10T01:29:57.641201778Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.641452 containerd[1507]: time="2026-03-10T01:29:57.641256002Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 10 01:29:57.641452 containerd[1507]: time="2026-03-10T01:29:57.641288322Z" level=info msg="NRI interface is disabled by configuration." Mar 10 01:29:57.641452 containerd[1507]: time="2026-03-10T01:29:57.641319164Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 10 01:29:57.642119 containerd[1507]: time="2026-03-10T01:29:57.641791306Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 10 01:29:57.642119 containerd[1507]: time="2026-03-10T01:29:57.641873620Z" level=info msg="Connect containerd service" Mar 10 01:29:57.642119 containerd[1507]: time="2026-03-10T01:29:57.641959145Z" level=info msg="using legacy CRI server" Mar 10 01:29:57.642119 containerd[1507]: time="2026-03-10T01:29:57.641975705Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 10 01:29:57.642442 containerd[1507]: time="2026-03-10T01:29:57.642183258Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 10 01:29:57.648587 containerd[1507]: time="2026-03-10T01:29:57.648534209Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 10 01:29:57.649387 containerd[1507]: time="2026-03-10T01:29:57.648711541Z" level=info msg="Start subscribing containerd event" Mar 10 01:29:57.649387 containerd[1507]: time="2026-03-10T01:29:57.648783362Z" level=info msg="Start recovering state" Mar 10 01:29:57.649387 containerd[1507]: time="2026-03-10T01:29:57.648897967Z" level=info msg="Start event monitor" Mar 10 01:29:57.649387 containerd[1507]: time="2026-03-10T01:29:57.648927733Z" level=info msg="Start snapshots syncer" Mar 10 01:29:57.649387 containerd[1507]: time="2026-03-10T01:29:57.648968558Z" level=info msg="Start cni network conf syncer for default" Mar 10 01:29:57.649387 containerd[1507]: time="2026-03-10T01:29:57.648985324Z" level=info msg="Start streaming server" Mar 10 01:29:57.649387 containerd[1507]: time="2026-03-10T01:29:57.649171750Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 10 01:29:57.651695 containerd[1507]: time="2026-03-10T01:29:57.651295170Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 10 01:29:57.651521 systemd[1]: Started containerd.service - containerd container runtime. Mar 10 01:29:57.651470 polkitd[1552]: Loading rules from directory /etc/polkit-1/rules.d Mar 10 01:29:57.651571 polkitd[1552]: Loading rules from directory /usr/share/polkit-1/rules.d Mar 10 01:29:57.654271 containerd[1507]: time="2026-03-10T01:29:57.652295892Z" level=info msg="containerd successfully booted in 0.097567s" Mar 10 01:29:57.657942 polkitd[1552]: Finished loading, compiling and executing 2 rules Mar 10 01:29:57.658750 dbus-daemon[1475]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Mar 10 01:29:57.659940 systemd[1]: Started polkit.service - Authorization Manager. Mar 10 01:29:57.663228 polkitd[1552]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Mar 10 01:29:57.688494 systemd-hostnamed[1517]: Hostname set to (static) Mar 10 01:29:57.734934 sshd_keygen[1504]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 10 01:29:57.765560 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 10 01:29:57.775659 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 10 01:29:57.785604 systemd[1]: Started sshd@0-10.230.50.50:22-68.220.241.50:33096.service - OpenSSH per-connection server daemon (68.220.241.50:33096). Mar 10 01:29:57.802386 systemd[1]: issuegen.service: Deactivated successfully. Mar 10 01:29:57.802719 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 10 01:29:57.809450 systemd-networkd[1419]: eth0: Gained IPv6LL Mar 10 01:29:57.814720 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 10 01:29:57.823816 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 10 01:29:57.827997 systemd[1]: Reached target network-online.target - Network is Online. Mar 10 01:29:57.840403 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:29:57.850804 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 10 01:29:57.857250 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 10 01:29:57.866853 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 10 01:29:57.873454 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 10 01:29:57.874633 systemd[1]: Reached target getty.target - Login Prompts. Mar 10 01:29:57.913508 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 10 01:29:58.224474 tar[1492]: linux-amd64/README.md Mar 10 01:29:58.239980 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 10 01:29:58.378362 sshd[1572]: Accepted publickey for core from 68.220.241.50 port 33096 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:29:58.380984 sshd[1572]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:29:58.401556 systemd-logind[1485]: New session 1 of user core. Mar 10 01:29:58.404100 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 10 01:29:58.412752 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 10 01:29:58.440695 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 10 01:29:58.454856 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 10 01:29:58.466584 (systemd)[1598]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 10 01:29:58.615644 systemd[1598]: Queued start job for default target default.target. Mar 10 01:29:58.633541 systemd[1598]: Created slice app.slice - User Application Slice. Mar 10 01:29:58.633804 systemd[1598]: Reached target paths.target - Paths. Mar 10 01:29:58.633933 systemd[1598]: Reached target timers.target - Timers. Mar 10 01:29:58.636422 systemd[1598]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 10 01:29:58.658097 systemd[1598]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 10 01:29:58.658305 systemd[1598]: Reached target sockets.target - Sockets. Mar 10 01:29:58.658330 systemd[1598]: Reached target basic.target - Basic System. Mar 10 01:29:58.658399 systemd[1598]: Reached target default.target - Main User Target. Mar 10 01:29:58.658464 systemd[1598]: Startup finished in 180ms. Mar 10 01:29:58.659338 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 10 01:29:58.676604 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 10 01:29:58.909015 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:29:58.919770 (kubelet)[1613]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 10 01:29:59.105642 systemd[1]: Started sshd@1-10.230.50.50:22-68.220.241.50:33106.service - OpenSSH per-connection server daemon (68.220.241.50:33106). Mar 10 01:29:59.316769 systemd-networkd[1419]: eth0: Ignoring DHCPv6 address 2a02:1348:179:8c8c:24:19ff:fee6:3232/128 (valid for 59min 59s, preferred for 59min 59s) which conflicts with 2a02:1348:179:8c8c:24:19ff:fee6:3232/64 assigned by NDisc. Mar 10 01:29:59.316781 systemd-networkd[1419]: eth0: Hint: use IPv6Token= setting to change the address generated by NDisc or set UseAutonomousPrefix=no. Mar 10 01:29:59.509343 kubelet[1613]: E0310 01:29:59.508462 1613 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 10 01:29:59.511738 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 10 01:29:59.512076 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 10 01:29:59.513725 systemd[1]: kubelet.service: Consumed 1.040s CPU time. Mar 10 01:29:59.671091 sshd[1620]: Accepted publickey for core from 68.220.241.50 port 33106 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:29:59.674431 sshd[1620]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:29:59.683447 systemd-logind[1485]: New session 2 of user core. Mar 10 01:29:59.696499 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 10 01:30:00.069137 sshd[1620]: pam_unix(sshd:session): session closed for user core Mar 10 01:30:00.074950 systemd[1]: sshd@1-10.230.50.50:22-68.220.241.50:33106.service: Deactivated successfully. Mar 10 01:30:00.077927 systemd[1]: session-2.scope: Deactivated successfully. Mar 10 01:30:00.080009 systemd-logind[1485]: Session 2 logged out. Waiting for processes to exit. Mar 10 01:30:00.081728 systemd-logind[1485]: Removed session 2. Mar 10 01:30:00.171653 systemd[1]: Started sshd@2-10.230.50.50:22-68.220.241.50:33108.service - OpenSSH per-connection server daemon (68.220.241.50:33108). Mar 10 01:30:00.738421 sshd[1631]: Accepted publickey for core from 68.220.241.50 port 33108 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:30:00.740892 sshd[1631]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:30:00.747621 systemd-logind[1485]: New session 3 of user core. Mar 10 01:30:00.757703 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 10 01:30:01.142568 sshd[1631]: pam_unix(sshd:session): session closed for user core Mar 10 01:30:01.148073 systemd-logind[1485]: Session 3 logged out. Waiting for processes to exit. Mar 10 01:30:01.148929 systemd[1]: sshd@2-10.230.50.50:22-68.220.241.50:33108.service: Deactivated successfully. Mar 10 01:30:01.152087 systemd[1]: session-3.scope: Deactivated successfully. Mar 10 01:30:01.155202 systemd-logind[1485]: Removed session 3. Mar 10 01:30:02.923009 login[1583]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Mar 10 01:30:02.934347 systemd-logind[1485]: New session 4 of user core. Mar 10 01:30:02.942606 login[1582]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Mar 10 01:30:02.948825 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 10 01:30:02.960463 systemd-logind[1485]: New session 5 of user core. Mar 10 01:30:02.972603 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 10 01:30:03.995267 coreos-metadata[1474]: Mar 10 01:30:03.994 WARN failed to locate config-drive, using the metadata service API instead Mar 10 01:30:04.020363 coreos-metadata[1474]: Mar 10 01:30:04.020 INFO Fetching http://169.254.169.254/openstack/2012-08-10/meta_data.json: Attempt #1 Mar 10 01:30:04.030530 coreos-metadata[1474]: Mar 10 01:30:04.030 INFO Fetch failed with 404: resource not found Mar 10 01:30:04.030627 coreos-metadata[1474]: Mar 10 01:30:04.030 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Mar 10 01:30:04.031666 coreos-metadata[1474]: Mar 10 01:30:04.031 INFO Fetch successful Mar 10 01:30:04.031858 coreos-metadata[1474]: Mar 10 01:30:04.031 INFO Fetching http://169.254.169.254/latest/meta-data/instance-id: Attempt #1 Mar 10 01:30:04.053414 coreos-metadata[1474]: Mar 10 01:30:04.053 INFO Fetch successful Mar 10 01:30:04.053667 coreos-metadata[1474]: Mar 10 01:30:04.053 INFO Fetching http://169.254.169.254/latest/meta-data/instance-type: Attempt #1 Mar 10 01:30:04.081364 coreos-metadata[1474]: Mar 10 01:30:04.081 INFO Fetch successful Mar 10 01:30:04.081709 coreos-metadata[1474]: Mar 10 01:30:04.081 INFO Fetching http://169.254.169.254/latest/meta-data/local-ipv4: Attempt #1 Mar 10 01:30:04.099658 coreos-metadata[1474]: Mar 10 01:30:04.099 INFO Fetch successful Mar 10 01:30:04.099975 coreos-metadata[1474]: Mar 10 01:30:04.099 INFO Fetching http://169.254.169.254/latest/meta-data/public-ipv4: Attempt #1 Mar 10 01:30:04.144502 coreos-metadata[1474]: Mar 10 01:30:04.144 INFO Fetch successful Mar 10 01:30:04.175450 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 10 01:30:04.177286 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 10 01:30:04.627267 coreos-metadata[1546]: Mar 10 01:30:04.624 WARN failed to locate config-drive, using the metadata service API instead Mar 10 01:30:04.647527 coreos-metadata[1546]: Mar 10 01:30:04.647 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys: Attempt #1 Mar 10 01:30:04.682056 coreos-metadata[1546]: Mar 10 01:30:04.682 INFO Fetch successful Mar 10 01:30:04.682407 coreos-metadata[1546]: Mar 10 01:30:04.682 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys/0/openssh-key: Attempt #1 Mar 10 01:30:04.716953 coreos-metadata[1546]: Mar 10 01:30:04.716 INFO Fetch successful Mar 10 01:30:04.718968 unknown[1546]: wrote ssh authorized keys file for user: core Mar 10 01:30:04.737928 update-ssh-keys[1671]: Updated "/home/core/.ssh/authorized_keys" Mar 10 01:30:04.738838 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Mar 10 01:30:04.741477 systemd[1]: Finished sshkeys.service. Mar 10 01:30:04.744512 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 10 01:30:04.749379 systemd[1]: Startup finished in 1.448s (kernel) + 14.361s (initrd) + 11.770s (userspace) = 27.580s. Mar 10 01:30:09.763378 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 10 01:30:09.779776 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:09.977722 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:09.980389 (kubelet)[1683]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 10 01:30:10.048184 kubelet[1683]: E0310 01:30:10.047736 1683 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 10 01:30:10.053429 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 10 01:30:10.053743 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 10 01:30:11.239438 systemd[1]: Started sshd@3-10.230.50.50:22-68.220.241.50:37226.service - OpenSSH per-connection server daemon (68.220.241.50:37226). Mar 10 01:30:11.802844 sshd[1691]: Accepted publickey for core from 68.220.241.50 port 37226 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:30:11.804969 sshd[1691]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:30:11.811788 systemd-logind[1485]: New session 6 of user core. Mar 10 01:30:11.824045 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 10 01:30:12.200031 sshd[1691]: pam_unix(sshd:session): session closed for user core Mar 10 01:30:12.205805 systemd[1]: sshd@3-10.230.50.50:22-68.220.241.50:37226.service: Deactivated successfully. Mar 10 01:30:12.207854 systemd[1]: session-6.scope: Deactivated successfully. Mar 10 01:30:12.208786 systemd-logind[1485]: Session 6 logged out. Waiting for processes to exit. Mar 10 01:30:12.210075 systemd-logind[1485]: Removed session 6. Mar 10 01:30:12.307640 systemd[1]: Started sshd@4-10.230.50.50:22-68.220.241.50:48298.service - OpenSSH per-connection server daemon (68.220.241.50:48298). Mar 10 01:30:12.855039 sshd[1698]: Accepted publickey for core from 68.220.241.50 port 48298 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:30:12.857085 sshd[1698]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:30:12.864039 systemd-logind[1485]: New session 7 of user core. Mar 10 01:30:12.877501 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 10 01:30:13.243163 sshd[1698]: pam_unix(sshd:session): session closed for user core Mar 10 01:30:13.248187 systemd[1]: sshd@4-10.230.50.50:22-68.220.241.50:48298.service: Deactivated successfully. Mar 10 01:30:13.250586 systemd[1]: session-7.scope: Deactivated successfully. Mar 10 01:30:13.251422 systemd-logind[1485]: Session 7 logged out. Waiting for processes to exit. Mar 10 01:30:13.253025 systemd-logind[1485]: Removed session 7. Mar 10 01:30:13.357711 systemd[1]: Started sshd@5-10.230.50.50:22-68.220.241.50:48304.service - OpenSSH per-connection server daemon (68.220.241.50:48304). Mar 10 01:30:13.929131 sshd[1705]: Accepted publickey for core from 68.220.241.50 port 48304 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:30:13.931174 sshd[1705]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:30:13.937654 systemd-logind[1485]: New session 8 of user core. Mar 10 01:30:13.947580 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 10 01:30:14.344329 sshd[1705]: pam_unix(sshd:session): session closed for user core Mar 10 01:30:14.349109 systemd-logind[1485]: Session 8 logged out. Waiting for processes to exit. Mar 10 01:30:14.349690 systemd[1]: sshd@5-10.230.50.50:22-68.220.241.50:48304.service: Deactivated successfully. Mar 10 01:30:14.352196 systemd[1]: session-8.scope: Deactivated successfully. Mar 10 01:30:14.356154 systemd-logind[1485]: Removed session 8. Mar 10 01:30:14.446625 systemd[1]: Started sshd@6-10.230.50.50:22-68.220.241.50:48308.service - OpenSSH per-connection server daemon (68.220.241.50:48308). Mar 10 01:30:15.019273 sshd[1712]: Accepted publickey for core from 68.220.241.50 port 48308 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:30:15.021550 sshd[1712]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:30:15.028028 systemd-logind[1485]: New session 9 of user core. Mar 10 01:30:15.036506 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 10 01:30:15.349531 sudo[1715]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 10 01:30:15.350021 sudo[1715]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 10 01:30:15.366542 sudo[1715]: pam_unix(sudo:session): session closed for user root Mar 10 01:30:15.455034 sshd[1712]: pam_unix(sshd:session): session closed for user core Mar 10 01:30:15.459383 systemd-logind[1485]: Session 9 logged out. Waiting for processes to exit. Mar 10 01:30:15.460610 systemd[1]: sshd@6-10.230.50.50:22-68.220.241.50:48308.service: Deactivated successfully. Mar 10 01:30:15.463104 systemd[1]: session-9.scope: Deactivated successfully. Mar 10 01:30:15.465875 systemd-logind[1485]: Removed session 9. Mar 10 01:30:15.551974 systemd[1]: Started sshd@7-10.230.50.50:22-68.220.241.50:48324.service - OpenSSH per-connection server daemon (68.220.241.50:48324). Mar 10 01:30:16.116329 sshd[1720]: Accepted publickey for core from 68.220.241.50 port 48324 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:30:16.117805 sshd[1720]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:30:16.125673 systemd-logind[1485]: New session 10 of user core. Mar 10 01:30:16.132513 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 10 01:30:16.437121 sudo[1724]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 10 01:30:16.438495 sudo[1724]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 10 01:30:16.443750 sudo[1724]: pam_unix(sudo:session): session closed for user root Mar 10 01:30:16.451743 sudo[1723]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Mar 10 01:30:16.452190 sudo[1723]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 10 01:30:16.474861 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Mar 10 01:30:16.477251 auditctl[1727]: No rules Mar 10 01:30:16.478243 systemd[1]: audit-rules.service: Deactivated successfully. Mar 10 01:30:16.478604 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Mar 10 01:30:16.481249 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 10 01:30:16.535035 augenrules[1745]: No rules Mar 10 01:30:16.536951 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 10 01:30:16.539519 sudo[1723]: pam_unix(sudo:session): session closed for user root Mar 10 01:30:16.643874 sshd[1720]: pam_unix(sshd:session): session closed for user core Mar 10 01:30:16.648940 systemd[1]: sshd@7-10.230.50.50:22-68.220.241.50:48324.service: Deactivated successfully. Mar 10 01:30:16.652004 systemd[1]: session-10.scope: Deactivated successfully. Mar 10 01:30:16.653999 systemd-logind[1485]: Session 10 logged out. Waiting for processes to exit. Mar 10 01:30:16.655410 systemd-logind[1485]: Removed session 10. Mar 10 01:30:16.759354 systemd[1]: Started sshd@8-10.230.50.50:22-68.220.241.50:48330.service - OpenSSH per-connection server daemon (68.220.241.50:48330). Mar 10 01:30:17.423050 sshd[1753]: Accepted publickey for core from 68.220.241.50 port 48330 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:30:17.425524 sshd[1753]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:30:17.432568 systemd-logind[1485]: New session 11 of user core. Mar 10 01:30:17.442588 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 10 01:30:17.747012 sudo[1756]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 10 01:30:17.747679 sudo[1756]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 10 01:30:18.232676 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 10 01:30:18.235127 (dockerd)[1773]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 10 01:30:18.670144 dockerd[1773]: time="2026-03-10T01:30:18.668726560Z" level=info msg="Starting up" Mar 10 01:30:18.841535 systemd[1]: var-lib-docker-metacopy\x2dcheck2110272884-merged.mount: Deactivated successfully. Mar 10 01:30:18.860887 dockerd[1773]: time="2026-03-10T01:30:18.860771794Z" level=info msg="Loading containers: start." Mar 10 01:30:19.003539 kernel: Initializing XFRM netlink socket Mar 10 01:30:19.126305 systemd-networkd[1419]: docker0: Link UP Mar 10 01:30:19.147214 dockerd[1773]: time="2026-03-10T01:30:19.147134824Z" level=info msg="Loading containers: done." Mar 10 01:30:19.169066 dockerd[1773]: time="2026-03-10T01:30:19.168994827Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 10 01:30:19.169345 dockerd[1773]: time="2026-03-10T01:30:19.169248323Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Mar 10 01:30:19.169543 dockerd[1773]: time="2026-03-10T01:30:19.169457342Z" level=info msg="Daemon has completed initialization" Mar 10 01:30:19.172703 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2341456923-merged.mount: Deactivated successfully. Mar 10 01:30:19.209930 dockerd[1773]: time="2026-03-10T01:30:19.209718311Z" level=info msg="API listen on /run/docker.sock" Mar 10 01:30:19.210095 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 10 01:30:19.941875 containerd[1507]: time="2026-03-10T01:30:19.940926892Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\"" Mar 10 01:30:20.304538 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 10 01:30:20.312532 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:20.576814 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:20.587756 (kubelet)[1924]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 10 01:30:20.658558 kubelet[1924]: E0310 01:30:20.658188 1924 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 10 01:30:20.660892 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 10 01:30:20.661135 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 10 01:30:20.933846 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2962169008.mount: Deactivated successfully. Mar 10 01:30:22.657466 containerd[1507]: time="2026-03-10T01:30:22.657200739Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:22.659120 containerd[1507]: time="2026-03-10T01:30:22.658982108Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.5: active requests=0, bytes read=27074505" Mar 10 01:30:22.660282 containerd[1507]: time="2026-03-10T01:30:22.660091092Z" level=info msg="ImageCreate event name:\"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:22.665256 containerd[1507]: time="2026-03-10T01:30:22.664260964Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:22.667399 containerd[1507]: time="2026-03-10T01:30:22.665998379Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.5\" with image id \"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\", size \"27071096\" in 2.724868672s" Mar 10 01:30:22.667399 containerd[1507]: time="2026-03-10T01:30:22.666062504Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\" returns image reference \"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\"" Mar 10 01:30:22.668943 containerd[1507]: time="2026-03-10T01:30:22.668901942Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\"" Mar 10 01:30:24.706538 containerd[1507]: time="2026-03-10T01:30:24.706336947Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:24.708383 containerd[1507]: time="2026-03-10T01:30:24.708195910Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.5: active requests=0, bytes read=21165831" Mar 10 01:30:24.709540 containerd[1507]: time="2026-03-10T01:30:24.708917133Z" level=info msg="ImageCreate event name:\"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:24.714346 containerd[1507]: time="2026-03-10T01:30:24.714295103Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:24.718252 containerd[1507]: time="2026-03-10T01:30:24.717149603Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.5\" with image id \"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\", size \"22822771\" in 2.048197858s" Mar 10 01:30:24.718252 containerd[1507]: time="2026-03-10T01:30:24.717213764Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\" returns image reference \"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\"" Mar 10 01:30:24.719854 containerd[1507]: time="2026-03-10T01:30:24.719812339Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\"" Mar 10 01:30:26.529260 containerd[1507]: time="2026-03-10T01:30:26.527912636Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:26.529260 containerd[1507]: time="2026-03-10T01:30:26.529152691Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.5: active requests=0, bytes read=15729832" Mar 10 01:30:26.530659 containerd[1507]: time="2026-03-10T01:30:26.530613637Z" level=info msg="ImageCreate event name:\"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:26.535057 containerd[1507]: time="2026-03-10T01:30:26.535004850Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:26.536692 containerd[1507]: time="2026-03-10T01:30:26.536648475Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.5\" with image id \"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\", size \"17386790\" in 1.816779828s" Mar 10 01:30:26.536775 containerd[1507]: time="2026-03-10T01:30:26.536723093Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\" returns image reference \"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\"" Mar 10 01:30:26.538779 containerd[1507]: time="2026-03-10T01:30:26.538728156Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\"" Mar 10 01:30:28.039986 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount249133066.mount: Deactivated successfully. Mar 10 01:30:28.627549 containerd[1507]: time="2026-03-10T01:30:28.626368863Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:28.629292 containerd[1507]: time="2026-03-10T01:30:28.629235870Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.5: active requests=0, bytes read=25861778" Mar 10 01:30:28.630238 containerd[1507]: time="2026-03-10T01:30:28.630142323Z" level=info msg="ImageCreate event name:\"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:28.634073 containerd[1507]: time="2026-03-10T01:30:28.633973075Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:28.636290 containerd[1507]: time="2026-03-10T01:30:28.635060092Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.5\" with image id \"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\", repo tag \"registry.k8s.io/kube-proxy:v1.34.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\", size \"25860789\" in 2.095940072s" Mar 10 01:30:28.636290 containerd[1507]: time="2026-03-10T01:30:28.635120496Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\" returns image reference \"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\"" Mar 10 01:30:28.637269 containerd[1507]: time="2026-03-10T01:30:28.637141044Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Mar 10 01:30:29.211433 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3149145553.mount: Deactivated successfully. Mar 10 01:30:29.373462 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Mar 10 01:30:30.834420 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 10 01:30:30.849277 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:31.001324 containerd[1507]: time="2026-03-10T01:30:30.999770038Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:31.047413 containerd[1507]: time="2026-03-10T01:30:31.047287988Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=22388015" Mar 10 01:30:31.127402 containerd[1507]: time="2026-03-10T01:30:31.125717108Z" level=info msg="ImageCreate event name:\"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:31.156937 containerd[1507]: time="2026-03-10T01:30:31.156857996Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:31.158889 containerd[1507]: time="2026-03-10T01:30:31.158842790Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"22384805\" in 2.521614172s" Mar 10 01:30:31.158979 containerd[1507]: time="2026-03-10T01:30:31.158930886Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\"" Mar 10 01:30:31.163056 containerd[1507]: time="2026-03-10T01:30:31.162943648Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Mar 10 01:30:31.218542 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:31.218810 (kubelet)[2070]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 10 01:30:31.286570 kubelet[2070]: E0310 01:30:31.286506 2070 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 10 01:30:31.290369 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 10 01:30:31.290649 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 10 01:30:31.959685 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3108612912.mount: Deactivated successfully. Mar 10 01:30:31.965908 containerd[1507]: time="2026-03-10T01:30:31.965828946Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:31.966906 containerd[1507]: time="2026-03-10T01:30:31.966797463Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=321226" Mar 10 01:30:31.967921 containerd[1507]: time="2026-03-10T01:30:31.967282117Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:31.971259 containerd[1507]: time="2026-03-10T01:30:31.970597633Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:31.972032 containerd[1507]: time="2026-03-10T01:30:31.971986734Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 808.963794ms" Mar 10 01:30:31.972143 containerd[1507]: time="2026-03-10T01:30:31.972069949Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Mar 10 01:30:31.973743 containerd[1507]: time="2026-03-10T01:30:31.973701660Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\"" Mar 10 01:30:32.900733 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3399644275.mount: Deactivated successfully. Mar 10 01:30:34.455304 containerd[1507]: time="2026-03-10T01:30:34.455133547Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.5-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:34.457141 containerd[1507]: time="2026-03-10T01:30:34.457025541Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.5-0: active requests=0, bytes read=22860682" Mar 10 01:30:34.459244 containerd[1507]: time="2026-03-10T01:30:34.457846778Z" level=info msg="ImageCreate event name:\"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:34.462138 containerd[1507]: time="2026-03-10T01:30:34.462096247Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:30:34.463983 containerd[1507]: time="2026-03-10T01:30:34.463947167Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.5-0\" with image id \"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\", repo tag \"registry.k8s.io/etcd:3.6.5-0\", repo digest \"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\", size \"22871747\" in 2.490192327s" Mar 10 01:30:34.464052 containerd[1507]: time="2026-03-10T01:30:34.463989327Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\" returns image reference \"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\"" Mar 10 01:30:39.736138 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:39.751532 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:39.791429 systemd[1]: Reloading requested from client PID 2169 ('systemctl') (unit session-11.scope)... Mar 10 01:30:39.791475 systemd[1]: Reloading... Mar 10 01:30:39.967293 zram_generator::config[2217]: No configuration found. Mar 10 01:30:40.145147 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 10 01:30:40.256998 systemd[1]: Reloading finished in 464 ms. Mar 10 01:30:40.343002 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:40.348739 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:40.351820 systemd[1]: kubelet.service: Deactivated successfully. Mar 10 01:30:40.352405 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:40.357612 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:40.518828 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:40.529918 (kubelet)[2277]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 10 01:30:40.633051 kubelet[2277]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 10 01:30:40.633700 kubelet[2277]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 10 01:30:40.634253 kubelet[2277]: I0310 01:30:40.633935 2277 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 10 01:30:41.438506 kubelet[2277]: I0310 01:30:41.438457 2277 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 10 01:30:41.438883 kubelet[2277]: I0310 01:30:41.438702 2277 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 10 01:30:41.441260 kubelet[2277]: I0310 01:30:41.440457 2277 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 10 01:30:41.441260 kubelet[2277]: I0310 01:30:41.440492 2277 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 10 01:30:41.441260 kubelet[2277]: I0310 01:30:41.440762 2277 server.go:956] "Client rotation is on, will bootstrap in background" Mar 10 01:30:41.448372 kubelet[2277]: E0310 01:30:41.448271 2277 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.230.50.50:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 10 01:30:41.449674 kubelet[2277]: I0310 01:30:41.449637 2277 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 10 01:30:41.456933 kubelet[2277]: E0310 01:30:41.456819 2277 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 10 01:30:41.456933 kubelet[2277]: I0310 01:30:41.456900 2277 server.go:1400] "CRI implementation should be updated to support RuntimeConfig. Falling back to using cgroupDriver from kubelet config." Mar 10 01:30:41.464196 kubelet[2277]: I0310 01:30:41.463646 2277 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 10 01:30:41.464792 kubelet[2277]: I0310 01:30:41.464738 2277 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 10 01:30:41.465246 kubelet[2277]: I0310 01:30:41.464888 2277 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"srv-eb8wl.gb1.brightbox.com","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 10 01:30:41.465598 kubelet[2277]: I0310 01:30:41.465576 2277 topology_manager.go:138] "Creating topology manager with none policy" Mar 10 01:30:41.466173 kubelet[2277]: I0310 01:30:41.465703 2277 container_manager_linux.go:306] "Creating device plugin manager" Mar 10 01:30:41.466173 kubelet[2277]: I0310 01:30:41.465880 2277 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 10 01:30:41.467706 kubelet[2277]: I0310 01:30:41.467682 2277 state_mem.go:36] "Initialized new in-memory state store" Mar 10 01:30:41.468135 kubelet[2277]: I0310 01:30:41.468114 2277 kubelet.go:475] "Attempting to sync node with API server" Mar 10 01:30:41.468308 kubelet[2277]: I0310 01:30:41.468287 2277 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 10 01:30:41.468486 kubelet[2277]: I0310 01:30:41.468466 2277 kubelet.go:387] "Adding apiserver pod source" Mar 10 01:30:41.468624 kubelet[2277]: I0310 01:30:41.468603 2277 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 10 01:30:41.470477 kubelet[2277]: E0310 01:30:41.469551 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.230.50.50:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-eb8wl.gb1.brightbox.com&limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 10 01:30:41.470864 kubelet[2277]: E0310 01:30:41.470831 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.230.50.50:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 10 01:30:41.471368 kubelet[2277]: I0310 01:30:41.471342 2277 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 10 01:30:41.472196 kubelet[2277]: I0310 01:30:41.472172 2277 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 10 01:30:41.472389 kubelet[2277]: I0310 01:30:41.472368 2277 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 10 01:30:41.472599 kubelet[2277]: W0310 01:30:41.472577 2277 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 10 01:30:41.478507 kubelet[2277]: I0310 01:30:41.478476 2277 server.go:1262] "Started kubelet" Mar 10 01:30:41.481971 kubelet[2277]: I0310 01:30:41.481947 2277 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 10 01:30:41.486659 kubelet[2277]: E0310 01:30:41.484609 2277 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.230.50.50:6443/api/v1/namespaces/default/events\": dial tcp 10.230.50.50:6443: connect: connection refused" event="&Event{ObjectMeta:{srv-eb8wl.gb1.brightbox.com.189b56bb7857eabc default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:srv-eb8wl.gb1.brightbox.com,UID:srv-eb8wl.gb1.brightbox.com,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:srv-eb8wl.gb1.brightbox.com,},FirstTimestamp:2026-03-10 01:30:41.478322876 +0000 UTC m=+0.941360347,LastTimestamp:2026-03-10 01:30:41.478322876 +0000 UTC m=+0.941360347,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:srv-eb8wl.gb1.brightbox.com,}" Mar 10 01:30:41.487349 kubelet[2277]: I0310 01:30:41.487312 2277 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 10 01:30:41.489678 kubelet[2277]: I0310 01:30:41.489636 2277 server.go:310] "Adding debug handlers to kubelet server" Mar 10 01:30:41.493146 kubelet[2277]: I0310 01:30:41.493058 2277 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 10 01:30:41.494482 kubelet[2277]: E0310 01:30:41.494455 2277 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" Mar 10 01:30:41.495643 kubelet[2277]: I0310 01:30:41.494918 2277 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 10 01:30:41.495643 kubelet[2277]: I0310 01:30:41.495021 2277 reconciler.go:29] "Reconciler: start to sync state" Mar 10 01:30:41.496429 kubelet[2277]: I0310 01:30:41.496018 2277 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 10 01:30:41.496429 kubelet[2277]: I0310 01:30:41.496128 2277 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 10 01:30:41.496536 kubelet[2277]: I0310 01:30:41.496432 2277 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 10 01:30:41.496716 kubelet[2277]: I0310 01:30:41.496692 2277 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 10 01:30:41.497998 kubelet[2277]: I0310 01:30:41.497976 2277 factory.go:223] Registration of the systemd container factory successfully Mar 10 01:30:41.498287 kubelet[2277]: I0310 01:30:41.498260 2277 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 10 01:30:41.498880 kubelet[2277]: E0310 01:30:41.498824 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.230.50.50:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 10 01:30:41.499189 kubelet[2277]: E0310 01:30:41.499131 2277 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.50.50:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-eb8wl.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.50.50:6443: connect: connection refused" interval="200ms" Mar 10 01:30:41.504654 kubelet[2277]: I0310 01:30:41.503783 2277 factory.go:223] Registration of the containerd container factory successfully Mar 10 01:30:41.504809 kubelet[2277]: E0310 01:30:41.504783 2277 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 10 01:30:41.536624 kubelet[2277]: I0310 01:30:41.536561 2277 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 10 01:30:41.539819 kubelet[2277]: I0310 01:30:41.539791 2277 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 10 01:30:41.539956 kubelet[2277]: I0310 01:30:41.539937 2277 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 10 01:30:41.540100 kubelet[2277]: I0310 01:30:41.540080 2277 kubelet.go:2428] "Starting kubelet main sync loop" Mar 10 01:30:41.540313 kubelet[2277]: E0310 01:30:41.540287 2277 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 10 01:30:41.545997 kubelet[2277]: I0310 01:30:41.545970 2277 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 10 01:30:41.545997 kubelet[2277]: I0310 01:30:41.545993 2277 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 10 01:30:41.546130 kubelet[2277]: I0310 01:30:41.546029 2277 state_mem.go:36] "Initialized new in-memory state store" Mar 10 01:30:41.548452 kubelet[2277]: E0310 01:30:41.548169 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.230.50.50:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 10 01:30:41.553273 kubelet[2277]: I0310 01:30:41.553251 2277 policy_none.go:49] "None policy: Start" Mar 10 01:30:41.553375 kubelet[2277]: I0310 01:30:41.553291 2277 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 10 01:30:41.553375 kubelet[2277]: I0310 01:30:41.553328 2277 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 10 01:30:41.554441 kubelet[2277]: I0310 01:30:41.554410 2277 policy_none.go:47] "Start" Mar 10 01:30:41.561153 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 10 01:30:41.580358 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 10 01:30:41.594567 kubelet[2277]: E0310 01:30:41.594536 2277 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" Mar 10 01:30:41.595469 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 10 01:30:41.598265 kubelet[2277]: E0310 01:30:41.597593 2277 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 10 01:30:41.598265 kubelet[2277]: I0310 01:30:41.597973 2277 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 10 01:30:41.598265 kubelet[2277]: I0310 01:30:41.597996 2277 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 10 01:30:41.600832 kubelet[2277]: I0310 01:30:41.600685 2277 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 10 01:30:41.601535 kubelet[2277]: E0310 01:30:41.601510 2277 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 10 01:30:41.601713 kubelet[2277]: E0310 01:30:41.601683 2277 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"srv-eb8wl.gb1.brightbox.com\" not found" Mar 10 01:30:41.656088 systemd[1]: Created slice kubepods-burstable-pod9b7ec1d61d2b4814f660e9039f4510ba.slice - libcontainer container kubepods-burstable-pod9b7ec1d61d2b4814f660e9039f4510ba.slice. Mar 10 01:30:41.670743 kubelet[2277]: E0310 01:30:41.670423 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.673925 systemd[1]: Created slice kubepods-burstable-pod1a5262e04781c5d88efb8d3004803b63.slice - libcontainer container kubepods-burstable-pod1a5262e04781c5d88efb8d3004803b63.slice. Mar 10 01:30:41.677547 kubelet[2277]: E0310 01:30:41.677524 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.682987 systemd[1]: Created slice kubepods-burstable-podf831c9ece0c738689b5872c661b399dc.slice - libcontainer container kubepods-burstable-podf831c9ece0c738689b5872c661b399dc.slice. Mar 10 01:30:41.685067 kubelet[2277]: E0310 01:30:41.685044 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.697260 kubelet[2277]: I0310 01:30:41.695424 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9b7ec1d61d2b4814f660e9039f4510ba-ca-certs\") pod \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" (UID: \"9b7ec1d61d2b4814f660e9039f4510ba\") " pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.700677 kubelet[2277]: I0310 01:30:41.700642 2277 kubelet_node_status.go:75] "Attempting to register node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.701191 kubelet[2277]: E0310 01:30:41.701162 2277 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.50.50:6443/api/v1/nodes\": dial tcp 10.230.50.50:6443: connect: connection refused" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.702640 kubelet[2277]: E0310 01:30:41.702595 2277 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.50.50:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-eb8wl.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.50.50:6443: connect: connection refused" interval="400ms" Mar 10 01:30:41.796605 kubelet[2277]: I0310 01:30:41.796270 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9b7ec1d61d2b4814f660e9039f4510ba-usr-share-ca-certificates\") pod \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" (UID: \"9b7ec1d61d2b4814f660e9039f4510ba\") " pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.796605 kubelet[2277]: I0310 01:30:41.796344 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-flexvolume-dir\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.796605 kubelet[2277]: I0310 01:30:41.796370 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-k8s-certs\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.796605 kubelet[2277]: I0310 01:30:41.796411 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-kubeconfig\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.796605 kubelet[2277]: I0310 01:30:41.796440 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f831c9ece0c738689b5872c661b399dc-kubeconfig\") pod \"kube-scheduler-srv-eb8wl.gb1.brightbox.com\" (UID: \"f831c9ece0c738689b5872c661b399dc\") " pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.797022 kubelet[2277]: I0310 01:30:41.796493 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9b7ec1d61d2b4814f660e9039f4510ba-k8s-certs\") pod \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" (UID: \"9b7ec1d61d2b4814f660e9039f4510ba\") " pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.797022 kubelet[2277]: I0310 01:30:41.796525 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-ca-certs\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.797022 kubelet[2277]: I0310 01:30:41.796551 2277 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-usr-share-ca-certificates\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.905457 kubelet[2277]: I0310 01:30:41.904930 2277 kubelet_node_status.go:75] "Attempting to register node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.905457 kubelet[2277]: E0310 01:30:41.905366 2277 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.50.50:6443/api/v1/nodes\": dial tcp 10.230.50.50:6443: connect: connection refused" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:41.975839 containerd[1507]: time="2026-03-10T01:30:41.975047860Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-srv-eb8wl.gb1.brightbox.com,Uid:9b7ec1d61d2b4814f660e9039f4510ba,Namespace:kube-system,Attempt:0,}" Mar 10 01:30:41.979843 containerd[1507]: time="2026-03-10T01:30:41.979493039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-srv-eb8wl.gb1.brightbox.com,Uid:1a5262e04781c5d88efb8d3004803b63,Namespace:kube-system,Attempt:0,}" Mar 10 01:30:41.987456 containerd[1507]: time="2026-03-10T01:30:41.987421395Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-srv-eb8wl.gb1.brightbox.com,Uid:f831c9ece0c738689b5872c661b399dc,Namespace:kube-system,Attempt:0,}" Mar 10 01:30:42.103780 kubelet[2277]: E0310 01:30:42.103699 2277 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.50.50:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-eb8wl.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.50.50:6443: connect: connection refused" interval="800ms" Mar 10 01:30:42.171190 update_engine[1486]: I20260310 01:30:42.171017 1486 update_attempter.cc:509] Updating boot flags... Mar 10 01:30:42.221526 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (2320) Mar 10 01:30:42.316341 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (2326) Mar 10 01:30:42.319093 kubelet[2277]: I0310 01:30:42.317073 2277 kubelet_node_status.go:75] "Attempting to register node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:42.319093 kubelet[2277]: E0310 01:30:42.317598 2277 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.50.50:6443/api/v1/nodes\": dial tcp 10.230.50.50:6443: connect: connection refused" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:42.387252 kubelet[2277]: E0310 01:30:42.387097 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.230.50.50:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 10 01:30:42.411036 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (2326) Mar 10 01:30:42.447834 kubelet[2277]: E0310 01:30:42.445944 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.230.50.50:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-eb8wl.gb1.brightbox.com&limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 10 01:30:42.524053 kubelet[2277]: E0310 01:30:42.523980 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.230.50.50:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 10 01:30:42.753544 kubelet[2277]: E0310 01:30:42.753471 2277 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.230.50.50:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 10 01:30:42.905507 kubelet[2277]: E0310 01:30:42.905403 2277 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.50.50:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-eb8wl.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.50.50:6443: connect: connection refused" interval="1.6s" Mar 10 01:30:43.032949 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2610172222.mount: Deactivated successfully. Mar 10 01:30:43.040988 containerd[1507]: time="2026-03-10T01:30:43.040929851Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 10 01:30:43.042622 containerd[1507]: time="2026-03-10T01:30:43.042489362Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 10 01:30:43.044244 containerd[1507]: time="2026-03-10T01:30:43.043320585Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 10 01:30:43.044548 containerd[1507]: time="2026-03-10T01:30:43.044513201Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 10 01:30:43.045634 containerd[1507]: time="2026-03-10T01:30:43.045517345Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312064" Mar 10 01:30:43.047240 containerd[1507]: time="2026-03-10T01:30:43.046400797Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 10 01:30:43.047240 containerd[1507]: time="2026-03-10T01:30:43.046499362Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 10 01:30:43.050245 containerd[1507]: time="2026-03-10T01:30:43.049465227Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 10 01:30:43.053025 containerd[1507]: time="2026-03-10T01:30:43.052986420Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.073418039s" Mar 10 01:30:43.055905 containerd[1507]: time="2026-03-10T01:30:43.055870566Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.068379768s" Mar 10 01:30:43.065937 containerd[1507]: time="2026-03-10T01:30:43.065874245Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.090620143s" Mar 10 01:30:43.122347 kubelet[2277]: I0310 01:30:43.121986 2277 kubelet_node_status.go:75] "Attempting to register node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:43.122542 kubelet[2277]: E0310 01:30:43.122394 2277 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.50.50:6443/api/v1/nodes\": dial tcp 10.230.50.50:6443: connect: connection refused" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:43.243317 containerd[1507]: time="2026-03-10T01:30:43.242578434Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:30:43.243317 containerd[1507]: time="2026-03-10T01:30:43.242819851Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:30:43.243317 containerd[1507]: time="2026-03-10T01:30:43.242898234Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:43.245672 containerd[1507]: time="2026-03-10T01:30:43.245353925Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:43.246729 containerd[1507]: time="2026-03-10T01:30:43.246642900Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:30:43.246807 containerd[1507]: time="2026-03-10T01:30:43.246748795Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:30:43.246884 containerd[1507]: time="2026-03-10T01:30:43.246801555Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:43.247012 containerd[1507]: time="2026-03-10T01:30:43.246967648Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:43.249270 containerd[1507]: time="2026-03-10T01:30:43.249121131Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:30:43.249270 containerd[1507]: time="2026-03-10T01:30:43.249191902Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:30:43.249270 containerd[1507]: time="2026-03-10T01:30:43.249212988Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:43.249613 containerd[1507]: time="2026-03-10T01:30:43.249561751Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:43.302131 systemd[1]: Started cri-containerd-3c76215f06f6c3c79ea1f999eb1a8df3f6a913646a964f76e25355ffe1c616b2.scope - libcontainer container 3c76215f06f6c3c79ea1f999eb1a8df3f6a913646a964f76e25355ffe1c616b2. Mar 10 01:30:43.306116 systemd[1]: Started cri-containerd-bc6d688dd53797112fd8685f75ce8a675e5d2b6615eff973d587be720b4f471a.scope - libcontainer container bc6d688dd53797112fd8685f75ce8a675e5d2b6615eff973d587be720b4f471a. Mar 10 01:30:43.315665 systemd[1]: Started cri-containerd-59ed78b12736976fb7844cab587716c7f8d84e3154e13b38c58e1b51f10b5cad.scope - libcontainer container 59ed78b12736976fb7844cab587716c7f8d84e3154e13b38c58e1b51f10b5cad. Mar 10 01:30:43.403178 containerd[1507]: time="2026-03-10T01:30:43.403011874Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-srv-eb8wl.gb1.brightbox.com,Uid:9b7ec1d61d2b4814f660e9039f4510ba,Namespace:kube-system,Attempt:0,} returns sandbox id \"bc6d688dd53797112fd8685f75ce8a675e5d2b6615eff973d587be720b4f471a\"" Mar 10 01:30:43.421613 containerd[1507]: time="2026-03-10T01:30:43.421491011Z" level=info msg="CreateContainer within sandbox \"bc6d688dd53797112fd8685f75ce8a675e5d2b6615eff973d587be720b4f471a\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 10 01:30:43.433647 containerd[1507]: time="2026-03-10T01:30:43.433395815Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-srv-eb8wl.gb1.brightbox.com,Uid:1a5262e04781c5d88efb8d3004803b63,Namespace:kube-system,Attempt:0,} returns sandbox id \"59ed78b12736976fb7844cab587716c7f8d84e3154e13b38c58e1b51f10b5cad\"" Mar 10 01:30:43.445408 containerd[1507]: time="2026-03-10T01:30:43.444699951Z" level=info msg="CreateContainer within sandbox \"59ed78b12736976fb7844cab587716c7f8d84e3154e13b38c58e1b51f10b5cad\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 10 01:30:43.445408 containerd[1507]: time="2026-03-10T01:30:43.444922821Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-srv-eb8wl.gb1.brightbox.com,Uid:f831c9ece0c738689b5872c661b399dc,Namespace:kube-system,Attempt:0,} returns sandbox id \"3c76215f06f6c3c79ea1f999eb1a8df3f6a913646a964f76e25355ffe1c616b2\"" Mar 10 01:30:43.450801 containerd[1507]: time="2026-03-10T01:30:43.450763640Z" level=info msg="CreateContainer within sandbox \"bc6d688dd53797112fd8685f75ce8a675e5d2b6615eff973d587be720b4f471a\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"fffa9f2aea7116a940178cf93582fdcdad1b949979a2dd8d12bb0337b1f6db0a\"" Mar 10 01:30:43.459465 containerd[1507]: time="2026-03-10T01:30:43.459384725Z" level=info msg="CreateContainer within sandbox \"59ed78b12736976fb7844cab587716c7f8d84e3154e13b38c58e1b51f10b5cad\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"a07054af50d855077c1eac3af57073ec032176808bb0d4cf7152f7568b2387f3\"" Mar 10 01:30:43.463607 containerd[1507]: time="2026-03-10T01:30:43.463573555Z" level=info msg="StartContainer for \"a07054af50d855077c1eac3af57073ec032176808bb0d4cf7152f7568b2387f3\"" Mar 10 01:30:43.465577 containerd[1507]: time="2026-03-10T01:30:43.465446443Z" level=info msg="StartContainer for \"fffa9f2aea7116a940178cf93582fdcdad1b949979a2dd8d12bb0337b1f6db0a\"" Mar 10 01:30:43.473491 containerd[1507]: time="2026-03-10T01:30:43.473297390Z" level=info msg="CreateContainer within sandbox \"3c76215f06f6c3c79ea1f999eb1a8df3f6a913646a964f76e25355ffe1c616b2\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 10 01:30:43.485057 containerd[1507]: time="2026-03-10T01:30:43.485016929Z" level=info msg="CreateContainer within sandbox \"3c76215f06f6c3c79ea1f999eb1a8df3f6a913646a964f76e25355ffe1c616b2\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f7fc0cf37dc3ba4b4b7c14c0f0a1433c0e1f0e2c225a27c5970ec7ea03dd8fa9\"" Mar 10 01:30:43.486234 containerd[1507]: time="2026-03-10T01:30:43.485845023Z" level=info msg="StartContainer for \"f7fc0cf37dc3ba4b4b7c14c0f0a1433c0e1f0e2c225a27c5970ec7ea03dd8fa9\"" Mar 10 01:30:43.514151 kubelet[2277]: E0310 01:30:43.514095 2277 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.230.50.50:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.230.50.50:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 10 01:30:43.529620 systemd[1]: Started cri-containerd-a07054af50d855077c1eac3af57073ec032176808bb0d4cf7152f7568b2387f3.scope - libcontainer container a07054af50d855077c1eac3af57073ec032176808bb0d4cf7152f7568b2387f3. Mar 10 01:30:43.549483 systemd[1]: Started cri-containerd-fffa9f2aea7116a940178cf93582fdcdad1b949979a2dd8d12bb0337b1f6db0a.scope - libcontainer container fffa9f2aea7116a940178cf93582fdcdad1b949979a2dd8d12bb0337b1f6db0a. Mar 10 01:30:43.568769 systemd[1]: Started cri-containerd-f7fc0cf37dc3ba4b4b7c14c0f0a1433c0e1f0e2c225a27c5970ec7ea03dd8fa9.scope - libcontainer container f7fc0cf37dc3ba4b4b7c14c0f0a1433c0e1f0e2c225a27c5970ec7ea03dd8fa9. Mar 10 01:30:43.661424 containerd[1507]: time="2026-03-10T01:30:43.661354680Z" level=info msg="StartContainer for \"fffa9f2aea7116a940178cf93582fdcdad1b949979a2dd8d12bb0337b1f6db0a\" returns successfully" Mar 10 01:30:43.673704 containerd[1507]: time="2026-03-10T01:30:43.673659787Z" level=info msg="StartContainer for \"a07054af50d855077c1eac3af57073ec032176808bb0d4cf7152f7568b2387f3\" returns successfully" Mar 10 01:30:43.712488 containerd[1507]: time="2026-03-10T01:30:43.712424085Z" level=info msg="StartContainer for \"f7fc0cf37dc3ba4b4b7c14c0f0a1433c0e1f0e2c225a27c5970ec7ea03dd8fa9\" returns successfully" Mar 10 01:30:44.596137 kubelet[2277]: E0310 01:30:44.596069 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:44.603171 kubelet[2277]: E0310 01:30:44.602675 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:44.607092 kubelet[2277]: E0310 01:30:44.606786 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:44.726769 kubelet[2277]: I0310 01:30:44.726719 2277 kubelet_node_status.go:75] "Attempting to register node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:45.613993 kubelet[2277]: E0310 01:30:45.613835 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:45.616387 kubelet[2277]: E0310 01:30:45.615109 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:45.616387 kubelet[2277]: E0310 01:30:45.615475 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.614185 kubelet[2277]: E0310 01:30:46.613880 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.614185 kubelet[2277]: E0310 01:30:46.614027 2277 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-eb8wl.gb1.brightbox.com\" not found" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.642436 kubelet[2277]: I0310 01:30:46.642360 2277 kubelet_node_status.go:78] "Successfully registered node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.695397 kubelet[2277]: I0310 01:30:46.695253 2277 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.715466 kubelet[2277]: E0310 01:30:46.715383 2277 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.715466 kubelet[2277]: I0310 01:30:46.715458 2277 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.721664 kubelet[2277]: E0310 01:30:46.721342 2277 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.721664 kubelet[2277]: I0310 01:30:46.721402 2277 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:46.729391 kubelet[2277]: E0310 01:30:46.729287 2277 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-srv-eb8wl.gb1.brightbox.com\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:47.488613 kubelet[2277]: I0310 01:30:47.486837 2277 apiserver.go:52] "Watching apiserver" Mar 10 01:30:47.495396 kubelet[2277]: I0310 01:30:47.495370 2277 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 10 01:30:47.613410 kubelet[2277]: I0310 01:30:47.612390 2277 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:47.623760 kubelet[2277]: I0310 01:30:47.623427 2277 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 10 01:30:47.845939 kubelet[2277]: I0310 01:30:47.845896 2277 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:47.853861 kubelet[2277]: I0310 01:30:47.853442 2277 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 10 01:30:49.093793 systemd[1]: Reloading requested from client PID 2580 ('systemctl') (unit session-11.scope)... Mar 10 01:30:49.093829 systemd[1]: Reloading... Mar 10 01:30:49.231300 zram_generator::config[2628]: No configuration found. Mar 10 01:30:49.410280 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 10 01:30:49.538991 systemd[1]: Reloading finished in 444 ms. Mar 10 01:30:49.595902 kubelet[2277]: I0310 01:30:49.595765 2277 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 10 01:30:49.596551 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:49.612940 systemd[1]: kubelet.service: Deactivated successfully. Mar 10 01:30:49.613634 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:49.613938 systemd[1]: kubelet.service: Consumed 1.517s CPU time, 121.4M memory peak, 0B memory swap peak. Mar 10 01:30:49.621587 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 10 01:30:49.853690 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 10 01:30:49.867766 (kubelet)[2683]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 10 01:30:49.996950 kubelet[2683]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 10 01:30:49.996950 kubelet[2683]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 10 01:30:49.999318 kubelet[2683]: I0310 01:30:49.997573 2683 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 10 01:30:50.018990 kubelet[2683]: I0310 01:30:50.018581 2683 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 10 01:30:50.018990 kubelet[2683]: I0310 01:30:50.018671 2683 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 10 01:30:50.024673 kubelet[2683]: I0310 01:30:50.024650 2683 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 10 01:30:50.024943 kubelet[2683]: I0310 01:30:50.024924 2683 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 10 01:30:50.025623 kubelet[2683]: I0310 01:30:50.025565 2683 server.go:956] "Client rotation is on, will bootstrap in background" Mar 10 01:30:50.028800 kubelet[2683]: I0310 01:30:50.028749 2683 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 10 01:30:50.040732 kubelet[2683]: I0310 01:30:50.040694 2683 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 10 01:30:50.064993 kubelet[2683]: E0310 01:30:50.064944 2683 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 10 01:30:50.065251 kubelet[2683]: I0310 01:30:50.065211 2683 server.go:1400] "CRI implementation should be updated to support RuntimeConfig. Falling back to using cgroupDriver from kubelet config." Mar 10 01:30:50.080798 kubelet[2683]: I0310 01:30:50.080753 2683 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 10 01:30:50.085474 kubelet[2683]: I0310 01:30:50.085375 2683 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 10 01:30:50.085840 kubelet[2683]: I0310 01:30:50.085595 2683 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"srv-eb8wl.gb1.brightbox.com","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 10 01:30:50.086168 kubelet[2683]: I0310 01:30:50.086147 2683 topology_manager.go:138] "Creating topology manager with none policy" Mar 10 01:30:50.086595 kubelet[2683]: I0310 01:30:50.086266 2683 container_manager_linux.go:306] "Creating device plugin manager" Mar 10 01:30:50.086595 kubelet[2683]: I0310 01:30:50.086343 2683 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 10 01:30:50.086883 kubelet[2683]: I0310 01:30:50.086865 2683 state_mem.go:36] "Initialized new in-memory state store" Mar 10 01:30:50.087534 kubelet[2683]: I0310 01:30:50.087512 2683 kubelet.go:475] "Attempting to sync node with API server" Mar 10 01:30:50.087685 kubelet[2683]: I0310 01:30:50.087644 2683 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 10 01:30:50.087970 kubelet[2683]: I0310 01:30:50.087951 2683 kubelet.go:387] "Adding apiserver pod source" Mar 10 01:30:50.089378 kubelet[2683]: I0310 01:30:50.089337 2683 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 10 01:30:50.119440 kubelet[2683]: I0310 01:30:50.118737 2683 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 10 01:30:50.119705 kubelet[2683]: I0310 01:30:50.119665 2683 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 10 01:30:50.119775 kubelet[2683]: I0310 01:30:50.119726 2683 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 10 01:30:50.130863 kubelet[2683]: I0310 01:30:50.130600 2683 server.go:1262] "Started kubelet" Mar 10 01:30:50.137100 kubelet[2683]: I0310 01:30:50.136790 2683 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 10 01:30:50.141955 kubelet[2683]: I0310 01:30:50.141736 2683 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 10 01:30:50.146603 kubelet[2683]: I0310 01:30:50.146331 2683 server.go:310] "Adding debug handlers to kubelet server" Mar 10 01:30:50.146700 kubelet[2683]: I0310 01:30:50.146619 2683 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 10 01:30:50.147839 kubelet[2683]: I0310 01:30:50.147135 2683 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 10 01:30:50.155456 kubelet[2683]: I0310 01:30:50.155422 2683 reconciler.go:29] "Reconciler: start to sync state" Mar 10 01:30:50.158243 kubelet[2683]: I0310 01:30:50.157437 2683 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 10 01:30:50.162145 kubelet[2683]: I0310 01:30:50.162092 2683 factory.go:223] Registration of the systemd container factory successfully Mar 10 01:30:50.162497 kubelet[2683]: I0310 01:30:50.162288 2683 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 10 01:30:50.165573 kubelet[2683]: I0310 01:30:50.164261 2683 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 10 01:30:50.165573 kubelet[2683]: I0310 01:30:50.165010 2683 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 10 01:30:50.168153 kubelet[2683]: E0310 01:30:50.168075 2683 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 10 01:30:50.177470 kubelet[2683]: I0310 01:30:50.177414 2683 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 10 01:30:50.182641 kubelet[2683]: I0310 01:30:50.182615 2683 factory.go:223] Registration of the containerd container factory successfully Mar 10 01:30:50.218351 kubelet[2683]: I0310 01:30:50.218070 2683 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 10 01:30:50.223911 kubelet[2683]: I0310 01:30:50.223484 2683 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 10 01:30:50.223911 kubelet[2683]: I0310 01:30:50.223531 2683 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 10 01:30:50.223911 kubelet[2683]: I0310 01:30:50.223571 2683 kubelet.go:2428] "Starting kubelet main sync loop" Mar 10 01:30:50.223911 kubelet[2683]: E0310 01:30:50.223657 2683 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 10 01:30:50.293777 kubelet[2683]: I0310 01:30:50.293724 2683 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 10 01:30:50.293777 kubelet[2683]: I0310 01:30:50.293755 2683 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 10 01:30:50.294103 kubelet[2683]: I0310 01:30:50.293831 2683 state_mem.go:36] "Initialized new in-memory state store" Mar 10 01:30:50.294103 kubelet[2683]: I0310 01:30:50.294074 2683 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 10 01:30:50.294259 kubelet[2683]: I0310 01:30:50.294098 2683 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 10 01:30:50.294259 kubelet[2683]: I0310 01:30:50.294154 2683 policy_none.go:49] "None policy: Start" Mar 10 01:30:50.294259 kubelet[2683]: I0310 01:30:50.294185 2683 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 10 01:30:50.294259 kubelet[2683]: I0310 01:30:50.294241 2683 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 10 01:30:50.294501 kubelet[2683]: I0310 01:30:50.294431 2683 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Mar 10 01:30:50.294501 kubelet[2683]: I0310 01:30:50.294452 2683 policy_none.go:47] "Start" Mar 10 01:30:50.309216 kubelet[2683]: E0310 01:30:50.308416 2683 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 10 01:30:50.309216 kubelet[2683]: I0310 01:30:50.308677 2683 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 10 01:30:50.309216 kubelet[2683]: I0310 01:30:50.308717 2683 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 10 01:30:50.318982 kubelet[2683]: E0310 01:30:50.318935 2683 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 10 01:30:50.323471 kubelet[2683]: I0310 01:30:50.319944 2683 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 10 01:30:50.325929 kubelet[2683]: I0310 01:30:50.325907 2683 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.332017 kubelet[2683]: I0310 01:30:50.329450 2683 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.332017 kubelet[2683]: I0310 01:30:50.329893 2683 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.350623 kubelet[2683]: I0310 01:30:50.350584 2683 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 10 01:30:50.352523 kubelet[2683]: E0310 01:30:50.351698 2683 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-srv-eb8wl.gb1.brightbox.com\" already exists" pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.352523 kubelet[2683]: I0310 01:30:50.350684 2683 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 10 01:30:50.352864 kubelet[2683]: E0310 01:30:50.351820 2683 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" already exists" pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.353237 kubelet[2683]: I0310 01:30:50.353199 2683 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 10 01:30:50.358407 kubelet[2683]: I0310 01:30:50.357518 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9b7ec1d61d2b4814f660e9039f4510ba-usr-share-ca-certificates\") pod \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" (UID: \"9b7ec1d61d2b4814f660e9039f4510ba\") " pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358407 kubelet[2683]: I0310 01:30:50.357564 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-ca-certs\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358407 kubelet[2683]: I0310 01:30:50.357597 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-k8s-certs\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358407 kubelet[2683]: I0310 01:30:50.357624 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-kubeconfig\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358407 kubelet[2683]: I0310 01:30:50.357652 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f831c9ece0c738689b5872c661b399dc-kubeconfig\") pod \"kube-scheduler-srv-eb8wl.gb1.brightbox.com\" (UID: \"f831c9ece0c738689b5872c661b399dc\") " pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358755 kubelet[2683]: I0310 01:30:50.357688 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9b7ec1d61d2b4814f660e9039f4510ba-k8s-certs\") pod \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" (UID: \"9b7ec1d61d2b4814f660e9039f4510ba\") " pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358755 kubelet[2683]: I0310 01:30:50.357716 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-flexvolume-dir\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358755 kubelet[2683]: I0310 01:30:50.357743 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1a5262e04781c5d88efb8d3004803b63-usr-share-ca-certificates\") pod \"kube-controller-manager-srv-eb8wl.gb1.brightbox.com\" (UID: \"1a5262e04781c5d88efb8d3004803b63\") " pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.358755 kubelet[2683]: I0310 01:30:50.357784 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9b7ec1d61d2b4814f660e9039f4510ba-ca-certs\") pod \"kube-apiserver-srv-eb8wl.gb1.brightbox.com\" (UID: \"9b7ec1d61d2b4814f660e9039f4510ba\") " pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.446799 kubelet[2683]: I0310 01:30:50.445440 2683 kubelet_node_status.go:75] "Attempting to register node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.471874 kubelet[2683]: I0310 01:30:50.471081 2683 kubelet_node_status.go:124] "Node was previously registered" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:50.471874 kubelet[2683]: I0310 01:30:50.471273 2683 kubelet_node_status.go:78] "Successfully registered node" node="srv-eb8wl.gb1.brightbox.com" Mar 10 01:30:51.120257 kubelet[2683]: I0310 01:30:51.119700 2683 apiserver.go:52] "Watching apiserver" Mar 10 01:30:51.147787 kubelet[2683]: I0310 01:30:51.147675 2683 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 10 01:30:51.314646 kubelet[2683]: I0310 01:30:51.314554 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-srv-eb8wl.gb1.brightbox.com" podStartSLOduration=4.314493998 podStartE2EDuration="4.314493998s" podCreationTimestamp="2026-03-10 01:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-10 01:30:51.299133807 +0000 UTC m=+1.406616928" watchObservedRunningTime="2026-03-10 01:30:51.314493998 +0000 UTC m=+1.421977112" Mar 10 01:30:51.327462 kubelet[2683]: I0310 01:30:51.327153 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-srv-eb8wl.gb1.brightbox.com" podStartSLOduration=4.327128383 podStartE2EDuration="4.327128383s" podCreationTimestamp="2026-03-10 01:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-10 01:30:51.314451312 +0000 UTC m=+1.421934419" watchObservedRunningTime="2026-03-10 01:30:51.327128383 +0000 UTC m=+1.434611497" Mar 10 01:30:51.356081 kubelet[2683]: I0310 01:30:51.355751 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-srv-eb8wl.gb1.brightbox.com" podStartSLOduration=1.355723882 podStartE2EDuration="1.355723882s" podCreationTimestamp="2026-03-10 01:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-10 01:30:51.328479836 +0000 UTC m=+1.435962955" watchObservedRunningTime="2026-03-10 01:30:51.355723882 +0000 UTC m=+1.463206997" Mar 10 01:30:54.934695 kubelet[2683]: I0310 01:30:54.933777 2683 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 10 01:30:54.936472 containerd[1507]: time="2026-03-10T01:30:54.936071043Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 10 01:30:54.937067 kubelet[2683]: I0310 01:30:54.936532 2683 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 10 01:30:55.832746 systemd[1]: Created slice kubepods-besteffort-pod65d26445_ab9e_4345_9f61_dc397168f2a6.slice - libcontainer container kubepods-besteffort-pod65d26445_ab9e_4345_9f61_dc397168f2a6.slice. Mar 10 01:30:55.897933 kubelet[2683]: I0310 01:30:55.897655 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwq6g\" (UniqueName: \"kubernetes.io/projected/65d26445-ab9e-4345-9f61-dc397168f2a6-kube-api-access-gwq6g\") pod \"kube-proxy-4fpnh\" (UID: \"65d26445-ab9e-4345-9f61-dc397168f2a6\") " pod="kube-system/kube-proxy-4fpnh" Mar 10 01:30:55.897933 kubelet[2683]: I0310 01:30:55.897732 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/65d26445-ab9e-4345-9f61-dc397168f2a6-kube-proxy\") pod \"kube-proxy-4fpnh\" (UID: \"65d26445-ab9e-4345-9f61-dc397168f2a6\") " pod="kube-system/kube-proxy-4fpnh" Mar 10 01:30:55.897933 kubelet[2683]: I0310 01:30:55.897778 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/65d26445-ab9e-4345-9f61-dc397168f2a6-xtables-lock\") pod \"kube-proxy-4fpnh\" (UID: \"65d26445-ab9e-4345-9f61-dc397168f2a6\") " pod="kube-system/kube-proxy-4fpnh" Mar 10 01:30:55.897933 kubelet[2683]: I0310 01:30:55.897818 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/65d26445-ab9e-4345-9f61-dc397168f2a6-lib-modules\") pod \"kube-proxy-4fpnh\" (UID: \"65d26445-ab9e-4345-9f61-dc397168f2a6\") " pod="kube-system/kube-proxy-4fpnh" Mar 10 01:30:56.132869 systemd[1]: Created slice kubepods-besteffort-pod3a6162c3_2133_4303_8a6b_e0104f5b2f2a.slice - libcontainer container kubepods-besteffort-pod3a6162c3_2133_4303_8a6b_e0104f5b2f2a.slice. Mar 10 01:30:56.142846 containerd[1507]: time="2026-03-10T01:30:56.142782446Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4fpnh,Uid:65d26445-ab9e-4345-9f61-dc397168f2a6,Namespace:kube-system,Attempt:0,}" Mar 10 01:30:56.182400 containerd[1507]: time="2026-03-10T01:30:56.182180854Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:30:56.182400 containerd[1507]: time="2026-03-10T01:30:56.182318927Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:30:56.182400 containerd[1507]: time="2026-03-10T01:30:56.182352812Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:56.183116 containerd[1507]: time="2026-03-10T01:30:56.182524795Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:56.201584 kubelet[2683]: I0310 01:30:56.201385 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3a6162c3-2133-4303-8a6b-e0104f5b2f2a-var-lib-calico\") pod \"tigera-operator-5588576f44-rgvw4\" (UID: \"3a6162c3-2133-4303-8a6b-e0104f5b2f2a\") " pod="tigera-operator/tigera-operator-5588576f44-rgvw4" Mar 10 01:30:56.201584 kubelet[2683]: I0310 01:30:56.201474 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t7pc\" (UniqueName: \"kubernetes.io/projected/3a6162c3-2133-4303-8a6b-e0104f5b2f2a-kube-api-access-8t7pc\") pod \"tigera-operator-5588576f44-rgvw4\" (UID: \"3a6162c3-2133-4303-8a6b-e0104f5b2f2a\") " pod="tigera-operator/tigera-operator-5588576f44-rgvw4" Mar 10 01:30:56.215847 systemd[1]: run-containerd-runc-k8s.io-9a97d75bd7dbf2dd2ed01b373dbd0a5b89c640836114ed9eb8b14127e5ef3d6b-runc.ApwwuZ.mount: Deactivated successfully. Mar 10 01:30:56.226414 systemd[1]: Started cri-containerd-9a97d75bd7dbf2dd2ed01b373dbd0a5b89c640836114ed9eb8b14127e5ef3d6b.scope - libcontainer container 9a97d75bd7dbf2dd2ed01b373dbd0a5b89c640836114ed9eb8b14127e5ef3d6b. Mar 10 01:30:56.273157 containerd[1507]: time="2026-03-10T01:30:56.272542968Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4fpnh,Uid:65d26445-ab9e-4345-9f61-dc397168f2a6,Namespace:kube-system,Attempt:0,} returns sandbox id \"9a97d75bd7dbf2dd2ed01b373dbd0a5b89c640836114ed9eb8b14127e5ef3d6b\"" Mar 10 01:30:56.283027 containerd[1507]: time="2026-03-10T01:30:56.282889478Z" level=info msg="CreateContainer within sandbox \"9a97d75bd7dbf2dd2ed01b373dbd0a5b89c640836114ed9eb8b14127e5ef3d6b\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 10 01:30:56.299081 containerd[1507]: time="2026-03-10T01:30:56.299048591Z" level=info msg="CreateContainer within sandbox \"9a97d75bd7dbf2dd2ed01b373dbd0a5b89c640836114ed9eb8b14127e5ef3d6b\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8591f00f7fe598166977d04d8792e50b21fd685daa7c55272711b09ab45b782a\"" Mar 10 01:30:56.300704 containerd[1507]: time="2026-03-10T01:30:56.300517806Z" level=info msg="StartContainer for \"8591f00f7fe598166977d04d8792e50b21fd685daa7c55272711b09ab45b782a\"" Mar 10 01:30:56.349331 systemd[1]: Started cri-containerd-8591f00f7fe598166977d04d8792e50b21fd685daa7c55272711b09ab45b782a.scope - libcontainer container 8591f00f7fe598166977d04d8792e50b21fd685daa7c55272711b09ab45b782a. Mar 10 01:30:56.398549 containerd[1507]: time="2026-03-10T01:30:56.398431650Z" level=info msg="StartContainer for \"8591f00f7fe598166977d04d8792e50b21fd685daa7c55272711b09ab45b782a\" returns successfully" Mar 10 01:30:56.440367 containerd[1507]: time="2026-03-10T01:30:56.439464100Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5588576f44-rgvw4,Uid:3a6162c3-2133-4303-8a6b-e0104f5b2f2a,Namespace:tigera-operator,Attempt:0,}" Mar 10 01:30:56.490273 containerd[1507]: time="2026-03-10T01:30:56.489913459Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:30:56.490273 containerd[1507]: time="2026-03-10T01:30:56.490012944Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:30:56.490273 containerd[1507]: time="2026-03-10T01:30:56.490029886Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:56.490927 containerd[1507]: time="2026-03-10T01:30:56.490784801Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:30:56.525491 systemd[1]: Started cri-containerd-c6788f27f696ea0399e0f357035e14566efe0996453bce13c8639469f63ae2d7.scope - libcontainer container c6788f27f696ea0399e0f357035e14566efe0996453bce13c8639469f63ae2d7. Mar 10 01:30:56.596908 containerd[1507]: time="2026-03-10T01:30:56.596814844Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5588576f44-rgvw4,Uid:3a6162c3-2133-4303-8a6b-e0104f5b2f2a,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"c6788f27f696ea0399e0f357035e14566efe0996453bce13c8639469f63ae2d7\"" Mar 10 01:30:56.600282 containerd[1507]: time="2026-03-10T01:30:56.600140946Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\"" Mar 10 01:30:57.316634 kubelet[2683]: I0310 01:30:57.316461 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-4fpnh" podStartSLOduration=2.316417579 podStartE2EDuration="2.316417579s" podCreationTimestamp="2026-03-10 01:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-10 01:30:57.302000409 +0000 UTC m=+7.409483519" watchObservedRunningTime="2026-03-10 01:30:57.316417579 +0000 UTC m=+7.423900694" Mar 10 01:30:58.502150 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1329837003.mount: Deactivated successfully. Mar 10 01:31:00.138251 containerd[1507]: time="2026-03-10T01:31:00.137929539Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.40.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:00.139494 containerd[1507]: time="2026-03-10T01:31:00.138972445Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.40.7: active requests=0, bytes read=40846156" Mar 10 01:31:00.140038 containerd[1507]: time="2026-03-10T01:31:00.140000417Z" level=info msg="ImageCreate event name:\"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:00.143168 containerd[1507]: time="2026-03-10T01:31:00.143118406Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:00.147049 containerd[1507]: time="2026-03-10T01:31:00.146978947Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.40.7\" with image id \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\", repo tag \"quay.io/tigera/operator:v1.40.7\", repo digest \"quay.io/tigera/operator@sha256:53260704fc6e638633b243729411222e01e1898647352a6e1a09cc046887973a\", size \"40842151\" in 3.546778513s" Mar 10 01:31:00.147049 containerd[1507]: time="2026-03-10T01:31:00.147031889Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.40.7\" returns image reference \"sha256:de04da31b5feb10fd313c39b7ac72d47ce9b5b8eb06161142e2e2283059a52c2\"" Mar 10 01:31:00.155261 containerd[1507]: time="2026-03-10T01:31:00.154997511Z" level=info msg="CreateContainer within sandbox \"c6788f27f696ea0399e0f357035e14566efe0996453bce13c8639469f63ae2d7\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 10 01:31:00.172700 containerd[1507]: time="2026-03-10T01:31:00.172556199Z" level=info msg="CreateContainer within sandbox \"c6788f27f696ea0399e0f357035e14566efe0996453bce13c8639469f63ae2d7\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"41f80b93dbd2a1def0b8987e1cd542169b26c27b42d3da4e5bb4252b014ab0bc\"" Mar 10 01:31:00.175758 containerd[1507]: time="2026-03-10T01:31:00.173566580Z" level=info msg="StartContainer for \"41f80b93dbd2a1def0b8987e1cd542169b26c27b42d3da4e5bb4252b014ab0bc\"" Mar 10 01:31:00.221997 systemd[1]: run-containerd-runc-k8s.io-41f80b93dbd2a1def0b8987e1cd542169b26c27b42d3da4e5bb4252b014ab0bc-runc.0x6vwr.mount: Deactivated successfully. Mar 10 01:31:00.234445 systemd[1]: Started cri-containerd-41f80b93dbd2a1def0b8987e1cd542169b26c27b42d3da4e5bb4252b014ab0bc.scope - libcontainer container 41f80b93dbd2a1def0b8987e1cd542169b26c27b42d3da4e5bb4252b014ab0bc. Mar 10 01:31:00.273860 containerd[1507]: time="2026-03-10T01:31:00.273666807Z" level=info msg="StartContainer for \"41f80b93dbd2a1def0b8987e1cd542169b26c27b42d3da4e5bb4252b014ab0bc\" returns successfully" Mar 10 01:31:00.318883 kubelet[2683]: I0310 01:31:00.318769 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-5588576f44-rgvw4" podStartSLOduration=0.76790448 podStartE2EDuration="4.31871213s" podCreationTimestamp="2026-03-10 01:30:56 +0000 UTC" firstStartedPulling="2026-03-10 01:30:56.599079947 +0000 UTC m=+6.706563055" lastFinishedPulling="2026-03-10 01:31:00.1498876 +0000 UTC m=+10.257370705" observedRunningTime="2026-03-10 01:31:00.317903515 +0000 UTC m=+10.425386635" watchObservedRunningTime="2026-03-10 01:31:00.31871213 +0000 UTC m=+10.426195250" Mar 10 01:31:08.049715 sudo[1756]: pam_unix(sudo:session): session closed for user root Mar 10 01:31:08.145412 sshd[1753]: pam_unix(sshd:session): session closed for user core Mar 10 01:31:08.157836 systemd[1]: sshd@8-10.230.50.50:22-68.220.241.50:48330.service: Deactivated successfully. Mar 10 01:31:08.163373 systemd[1]: session-11.scope: Deactivated successfully. Mar 10 01:31:08.163965 systemd[1]: session-11.scope: Consumed 8.411s CPU time, 157.4M memory peak, 0B memory swap peak. Mar 10 01:31:08.168597 systemd-logind[1485]: Session 11 logged out. Waiting for processes to exit. Mar 10 01:31:08.171527 systemd-logind[1485]: Removed session 11. Mar 10 01:31:12.447297 systemd[1]: Created slice kubepods-besteffort-pod51c5114e_9642_4a30_8ebb_f32b5f920061.slice - libcontainer container kubepods-besteffort-pod51c5114e_9642_4a30_8ebb_f32b5f920061.slice. Mar 10 01:31:12.528666 kubelet[2683]: I0310 01:31:12.528345 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51c5114e-9642-4a30-8ebb-f32b5f920061-tigera-ca-bundle\") pod \"calico-typha-6787995c7-7f9vg\" (UID: \"51c5114e-9642-4a30-8ebb-f32b5f920061\") " pod="calico-system/calico-typha-6787995c7-7f9vg" Mar 10 01:31:12.528666 kubelet[2683]: I0310 01:31:12.528435 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x7q5\" (UniqueName: \"kubernetes.io/projected/51c5114e-9642-4a30-8ebb-f32b5f920061-kube-api-access-4x7q5\") pod \"calico-typha-6787995c7-7f9vg\" (UID: \"51c5114e-9642-4a30-8ebb-f32b5f920061\") " pod="calico-system/calico-typha-6787995c7-7f9vg" Mar 10 01:31:12.528666 kubelet[2683]: I0310 01:31:12.528491 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/51c5114e-9642-4a30-8ebb-f32b5f920061-typha-certs\") pod \"calico-typha-6787995c7-7f9vg\" (UID: \"51c5114e-9642-4a30-8ebb-f32b5f920061\") " pod="calico-system/calico-typha-6787995c7-7f9vg" Mar 10 01:31:12.566891 systemd[1]: Created slice kubepods-besteffort-podbd178a7f_8866_4467_8b65_479d48e91fb8.slice - libcontainer container kubepods-besteffort-podbd178a7f_8866_4467_8b65_479d48e91fb8.slice. Mar 10 01:31:12.629352 kubelet[2683]: I0310 01:31:12.629290 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-var-lib-calico\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.631391 kubelet[2683]: I0310 01:31:12.631299 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-lib-modules\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633274 kubelet[2683]: I0310 01:31:12.631539 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-cni-net-dir\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633274 kubelet[2683]: I0310 01:31:12.631584 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/bd178a7f-8866-4467-8b65-479d48e91fb8-node-certs\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633274 kubelet[2683]: I0310 01:31:12.631630 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-cni-bin-dir\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633274 kubelet[2683]: I0310 01:31:12.631668 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-cni-log-dir\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633274 kubelet[2683]: I0310 01:31:12.631695 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd178a7f-8866-4467-8b65-479d48e91fb8-tigera-ca-bundle\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633529 kubelet[2683]: I0310 01:31:12.631722 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-var-run-calico\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633529 kubelet[2683]: I0310 01:31:12.631926 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nodeproc\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-nodeproc\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633529 kubelet[2683]: I0310 01:31:12.632007 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-policysync\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633529 kubelet[2683]: I0310 01:31:12.632100 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n449\" (UniqueName: \"kubernetes.io/projected/bd178a7f-8866-4467-8b65-479d48e91fb8-kube-api-access-9n449\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633529 kubelet[2683]: I0310 01:31:12.632197 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-sys-fs\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633762 kubelet[2683]: I0310 01:31:12.632285 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpffs\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-bpffs\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633762 kubelet[2683]: I0310 01:31:12.632363 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-flexvol-driver-host\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.633762 kubelet[2683]: I0310 01:31:12.632397 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/bd178a7f-8866-4467-8b65-479d48e91fb8-xtables-lock\") pod \"calico-node-7sq4w\" (UID: \"bd178a7f-8866-4467-8b65-479d48e91fb8\") " pod="calico-system/calico-node-7sq4w" Mar 10 01:31:12.700096 kubelet[2683]: E0310 01:31:12.699833 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:12.733799 kubelet[2683]: I0310 01:31:12.733692 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr69w\" (UniqueName: \"kubernetes.io/projected/71ea22a7-e0cb-4bd4-871f-854f4db6a991-kube-api-access-tr69w\") pod \"csi-node-driver-fhcl2\" (UID: \"71ea22a7-e0cb-4bd4-871f-854f4db6a991\") " pod="calico-system/csi-node-driver-fhcl2" Mar 10 01:31:12.735519 kubelet[2683]: I0310 01:31:12.735484 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/71ea22a7-e0cb-4bd4-871f-854f4db6a991-registration-dir\") pod \"csi-node-driver-fhcl2\" (UID: \"71ea22a7-e0cb-4bd4-871f-854f4db6a991\") " pod="calico-system/csi-node-driver-fhcl2" Mar 10 01:31:12.736997 kubelet[2683]: E0310 01:31:12.736834 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.736997 kubelet[2683]: W0310 01:31:12.736893 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.736997 kubelet[2683]: E0310 01:31:12.736966 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.738287 kubelet[2683]: E0310 01:31:12.738263 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.738287 kubelet[2683]: W0310 01:31:12.738283 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.738427 kubelet[2683]: E0310 01:31:12.738299 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.738900 kubelet[2683]: I0310 01:31:12.738549 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/71ea22a7-e0cb-4bd4-871f-854f4db6a991-varrun\") pod \"csi-node-driver-fhcl2\" (UID: \"71ea22a7-e0cb-4bd4-871f-854f4db6a991\") " pod="calico-system/csi-node-driver-fhcl2" Mar 10 01:31:12.740026 kubelet[2683]: E0310 01:31:12.739294 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.740238 kubelet[2683]: W0310 01:31:12.740146 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.740238 kubelet[2683]: E0310 01:31:12.740174 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.740776 kubelet[2683]: E0310 01:31:12.740656 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.740776 kubelet[2683]: W0310 01:31:12.740674 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.740776 kubelet[2683]: E0310 01:31:12.740690 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.741329 kubelet[2683]: E0310 01:31:12.741274 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.741329 kubelet[2683]: W0310 01:31:12.741293 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.741329 kubelet[2683]: E0310 01:31:12.741309 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.741904 kubelet[2683]: E0310 01:31:12.741783 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.741904 kubelet[2683]: W0310 01:31:12.741801 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.741904 kubelet[2683]: E0310 01:31:12.741816 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.742441 kubelet[2683]: E0310 01:31:12.742322 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.742441 kubelet[2683]: W0310 01:31:12.742339 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.742441 kubelet[2683]: E0310 01:31:12.742355 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.742937 kubelet[2683]: E0310 01:31:12.742820 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.742937 kubelet[2683]: W0310 01:31:12.742837 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.742937 kubelet[2683]: E0310 01:31:12.742852 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.745709 kubelet[2683]: E0310 01:31:12.745573 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.745709 kubelet[2683]: W0310 01:31:12.745592 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.745709 kubelet[2683]: E0310 01:31:12.745609 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.746460 kubelet[2683]: E0310 01:31:12.746406 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.746460 kubelet[2683]: W0310 01:31:12.746424 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.746460 kubelet[2683]: E0310 01:31:12.746439 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.749512 kubelet[2683]: E0310 01:31:12.749340 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.749512 kubelet[2683]: W0310 01:31:12.749362 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.749512 kubelet[2683]: E0310 01:31:12.749390 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.750035 kubelet[2683]: E0310 01:31:12.749890 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.750035 kubelet[2683]: W0310 01:31:12.749907 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.750035 kubelet[2683]: E0310 01:31:12.749922 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.751124 kubelet[2683]: E0310 01:31:12.751105 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.751211 kubelet[2683]: W0310 01:31:12.751192 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.751354 kubelet[2683]: E0310 01:31:12.751334 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.751725 kubelet[2683]: E0310 01:31:12.751706 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.751831 kubelet[2683]: W0310 01:31:12.751811 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.751940 kubelet[2683]: E0310 01:31:12.751921 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.752492 kubelet[2683]: E0310 01:31:12.752350 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.752492 kubelet[2683]: W0310 01:31:12.752367 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.752492 kubelet[2683]: E0310 01:31:12.752382 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.753307 kubelet[2683]: E0310 01:31:12.752968 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.753307 kubelet[2683]: W0310 01:31:12.752985 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.753307 kubelet[2683]: E0310 01:31:12.753000 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.753652 kubelet[2683]: E0310 01:31:12.753634 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.753870 kubelet[2683]: W0310 01:31:12.753748 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.753870 kubelet[2683]: E0310 01:31:12.753772 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.755316 kubelet[2683]: E0310 01:31:12.754253 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.755316 kubelet[2683]: W0310 01:31:12.754276 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.755316 kubelet[2683]: E0310 01:31:12.754291 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.758338 kubelet[2683]: E0310 01:31:12.758139 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.758338 kubelet[2683]: W0310 01:31:12.758164 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.758338 kubelet[2683]: E0310 01:31:12.758187 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.758788 kubelet[2683]: E0310 01:31:12.758623 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.758788 kubelet[2683]: W0310 01:31:12.758640 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.758788 kubelet[2683]: E0310 01:31:12.758656 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.760823 kubelet[2683]: E0310 01:31:12.760169 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.760823 kubelet[2683]: W0310 01:31:12.760188 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.760823 kubelet[2683]: E0310 01:31:12.760203 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.762264 kubelet[2683]: E0310 01:31:12.761859 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.762264 kubelet[2683]: W0310 01:31:12.761878 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.762264 kubelet[2683]: E0310 01:31:12.761894 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.764777 containerd[1507]: time="2026-03-10T01:31:12.764613014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6787995c7-7f9vg,Uid:51c5114e-9642-4a30-8ebb-f32b5f920061,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:12.766816 kubelet[2683]: E0310 01:31:12.765504 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.766816 kubelet[2683]: W0310 01:31:12.765525 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.766816 kubelet[2683]: E0310 01:31:12.765554 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.766816 kubelet[2683]: E0310 01:31:12.766531 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.766816 kubelet[2683]: W0310 01:31:12.766545 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.766816 kubelet[2683]: E0310 01:31:12.766561 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.768811 kubelet[2683]: E0310 01:31:12.767857 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.768811 kubelet[2683]: W0310 01:31:12.767875 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.768811 kubelet[2683]: E0310 01:31:12.767891 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.769539 kubelet[2683]: E0310 01:31:12.769196 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.769539 kubelet[2683]: W0310 01:31:12.769258 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.769539 kubelet[2683]: E0310 01:31:12.769282 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.774827 kubelet[2683]: E0310 01:31:12.774287 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.774827 kubelet[2683]: W0310 01:31:12.774308 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.774827 kubelet[2683]: E0310 01:31:12.774563 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.776421 kubelet[2683]: E0310 01:31:12.775990 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.776421 kubelet[2683]: W0310 01:31:12.776008 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.776421 kubelet[2683]: E0310 01:31:12.776024 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.778667 kubelet[2683]: E0310 01:31:12.778307 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.778667 kubelet[2683]: W0310 01:31:12.778327 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.778667 kubelet[2683]: E0310 01:31:12.778345 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.779789 kubelet[2683]: E0310 01:31:12.779131 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.779789 kubelet[2683]: W0310 01:31:12.779149 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.779789 kubelet[2683]: E0310 01:31:12.779164 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.780949 kubelet[2683]: E0310 01:31:12.780878 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.780949 kubelet[2683]: W0310 01:31:12.780896 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.780949 kubelet[2683]: E0310 01:31:12.780911 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.783259 kubelet[2683]: E0310 01:31:12.782433 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.783259 kubelet[2683]: W0310 01:31:12.782452 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.783259 kubelet[2683]: E0310 01:31:12.782492 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.786491 kubelet[2683]: E0310 01:31:12.786202 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.786491 kubelet[2683]: W0310 01:31:12.786238 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.786491 kubelet[2683]: E0310 01:31:12.786261 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.787291 kubelet[2683]: E0310 01:31:12.786869 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.787291 kubelet[2683]: W0310 01:31:12.786892 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.787291 kubelet[2683]: E0310 01:31:12.786908 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.787697 kubelet[2683]: E0310 01:31:12.787602 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.787697 kubelet[2683]: W0310 01:31:12.787619 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.787697 kubelet[2683]: E0310 01:31:12.787634 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.790650 kubelet[2683]: E0310 01:31:12.790513 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.790650 kubelet[2683]: W0310 01:31:12.790529 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.790650 kubelet[2683]: E0310 01:31:12.790556 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.794103 kubelet[2683]: I0310 01:31:12.793926 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71ea22a7-e0cb-4bd4-871f-854f4db6a991-kubelet-dir\") pod \"csi-node-driver-fhcl2\" (UID: \"71ea22a7-e0cb-4bd4-871f-854f4db6a991\") " pod="calico-system/csi-node-driver-fhcl2" Mar 10 01:31:12.794700 kubelet[2683]: E0310 01:31:12.794291 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.794700 kubelet[2683]: W0310 01:31:12.794305 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.794700 kubelet[2683]: E0310 01:31:12.794321 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.796295 kubelet[2683]: E0310 01:31:12.795164 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.796295 kubelet[2683]: W0310 01:31:12.795182 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.796295 kubelet[2683]: E0310 01:31:12.795197 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.796663 kubelet[2683]: E0310 01:31:12.796479 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.796663 kubelet[2683]: W0310 01:31:12.796493 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.796663 kubelet[2683]: E0310 01:31:12.796508 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.800737 kubelet[2683]: E0310 01:31:12.799457 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.800737 kubelet[2683]: W0310 01:31:12.799474 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.800737 kubelet[2683]: E0310 01:31:12.799500 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.800737 kubelet[2683]: I0310 01:31:12.800380 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/71ea22a7-e0cb-4bd4-871f-854f4db6a991-socket-dir\") pod \"csi-node-driver-fhcl2\" (UID: \"71ea22a7-e0cb-4bd4-871f-854f4db6a991\") " pod="calico-system/csi-node-driver-fhcl2" Mar 10 01:31:12.804501 kubelet[2683]: E0310 01:31:12.804289 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.804501 kubelet[2683]: W0310 01:31:12.804309 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.804501 kubelet[2683]: E0310 01:31:12.804325 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.804894 kubelet[2683]: E0310 01:31:12.804686 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.804894 kubelet[2683]: W0310 01:31:12.804699 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.804894 kubelet[2683]: E0310 01:31:12.804712 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.806566 kubelet[2683]: E0310 01:31:12.805344 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.806566 kubelet[2683]: W0310 01:31:12.805361 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.806566 kubelet[2683]: E0310 01:31:12.805376 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.809428 kubelet[2683]: E0310 01:31:12.809204 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.809428 kubelet[2683]: W0310 01:31:12.809237 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.809428 kubelet[2683]: E0310 01:31:12.809256 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.811987 kubelet[2683]: E0310 01:31:12.811030 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.811987 kubelet[2683]: W0310 01:31:12.811066 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.811987 kubelet[2683]: E0310 01:31:12.811085 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.813594 kubelet[2683]: E0310 01:31:12.813205 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.813594 kubelet[2683]: W0310 01:31:12.813281 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.813594 kubelet[2683]: E0310 01:31:12.813301 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.816903 kubelet[2683]: E0310 01:31:12.816351 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.816903 kubelet[2683]: W0310 01:31:12.816372 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.816903 kubelet[2683]: E0310 01:31:12.816399 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.816903 kubelet[2683]: E0310 01:31:12.816670 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.816903 kubelet[2683]: W0310 01:31:12.816683 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.816903 kubelet[2683]: E0310 01:31:12.816697 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.853244 kubelet[2683]: E0310 01:31:12.851824 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.853244 kubelet[2683]: W0310 01:31:12.851854 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.853244 kubelet[2683]: E0310 01:31:12.851894 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.878304 containerd[1507]: time="2026-03-10T01:31:12.878253366Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-7sq4w,Uid:bd178a7f-8866-4467-8b65-479d48e91fb8,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:12.902807 kubelet[2683]: E0310 01:31:12.902770 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.902898 kubelet[2683]: W0310 01:31:12.902819 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.902898 kubelet[2683]: E0310 01:31:12.902858 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.904256 kubelet[2683]: E0310 01:31:12.903795 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.904256 kubelet[2683]: W0310 01:31:12.903816 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.904256 kubelet[2683]: E0310 01:31:12.903832 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.904911 kubelet[2683]: E0310 01:31:12.904884 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.904911 kubelet[2683]: W0310 01:31:12.904904 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.905060 kubelet[2683]: E0310 01:31:12.904921 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.905744 kubelet[2683]: E0310 01:31:12.905700 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.905868 kubelet[2683]: W0310 01:31:12.905731 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.905963 kubelet[2683]: E0310 01:31:12.905876 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.906726 kubelet[2683]: E0310 01:31:12.906701 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.906726 kubelet[2683]: W0310 01:31:12.906727 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.907273 kubelet[2683]: E0310 01:31:12.907211 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.907714 kubelet[2683]: E0310 01:31:12.907686 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.907714 kubelet[2683]: W0310 01:31:12.907705 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.907921 kubelet[2683]: E0310 01:31:12.907735 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.908377 containerd[1507]: time="2026-03-10T01:31:12.908170535Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:12.908377 containerd[1507]: time="2026-03-10T01:31:12.908340236Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:12.908646 containerd[1507]: time="2026-03-10T01:31:12.908364796Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:12.908743 containerd[1507]: time="2026-03-10T01:31:12.908619678Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:12.909185 kubelet[2683]: E0310 01:31:12.909162 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.909185 kubelet[2683]: W0310 01:31:12.909183 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.910347 kubelet[2683]: E0310 01:31:12.909200 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.910347 kubelet[2683]: E0310 01:31:12.909523 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.910347 kubelet[2683]: W0310 01:31:12.909537 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.910347 kubelet[2683]: E0310 01:31:12.909551 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.910347 kubelet[2683]: E0310 01:31:12.909918 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.910347 kubelet[2683]: W0310 01:31:12.909932 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.910347 kubelet[2683]: E0310 01:31:12.909955 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.910347 kubelet[2683]: E0310 01:31:12.910279 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.910347 kubelet[2683]: W0310 01:31:12.910293 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.910347 kubelet[2683]: E0310 01:31:12.910307 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.911354 kubelet[2683]: E0310 01:31:12.911286 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.911354 kubelet[2683]: W0310 01:31:12.911306 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.911354 kubelet[2683]: E0310 01:31:12.911353 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.911812 kubelet[2683]: E0310 01:31:12.911788 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.911812 kubelet[2683]: W0310 01:31:12.911809 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.911951 kubelet[2683]: E0310 01:31:12.911825 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.912349 kubelet[2683]: E0310 01:31:12.912318 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.912449 kubelet[2683]: W0310 01:31:12.912344 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.912539 kubelet[2683]: E0310 01:31:12.912479 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.913381 kubelet[2683]: E0310 01:31:12.913296 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.913381 kubelet[2683]: W0310 01:31:12.913317 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.913514 kubelet[2683]: E0310 01:31:12.913467 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.914260 kubelet[2683]: E0310 01:31:12.914128 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.914260 kubelet[2683]: W0310 01:31:12.914148 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.914385 kubelet[2683]: E0310 01:31:12.914285 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.915080 kubelet[2683]: E0310 01:31:12.914975 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.915080 kubelet[2683]: W0310 01:31:12.914994 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.915080 kubelet[2683]: E0310 01:31:12.915074 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.917345 kubelet[2683]: E0310 01:31:12.917299 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.917345 kubelet[2683]: W0310 01:31:12.917342 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.917497 kubelet[2683]: E0310 01:31:12.917358 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.918145 kubelet[2683]: E0310 01:31:12.918114 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.918145 kubelet[2683]: W0310 01:31:12.918135 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.918270 kubelet[2683]: E0310 01:31:12.918152 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.921171 kubelet[2683]: E0310 01:31:12.921137 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.921171 kubelet[2683]: W0310 01:31:12.921168 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.921484 kubelet[2683]: E0310 01:31:12.921185 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.922567 kubelet[2683]: E0310 01:31:12.922504 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.922567 kubelet[2683]: W0310 01:31:12.922524 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.924829 kubelet[2683]: E0310 01:31:12.923231 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.924829 kubelet[2683]: E0310 01:31:12.924111 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.924829 kubelet[2683]: W0310 01:31:12.924126 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.924829 kubelet[2683]: E0310 01:31:12.924168 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.924829 kubelet[2683]: E0310 01:31:12.924572 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.924829 kubelet[2683]: W0310 01:31:12.924597 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.924829 kubelet[2683]: E0310 01:31:12.924646 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.925944 kubelet[2683]: E0310 01:31:12.925770 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.925944 kubelet[2683]: W0310 01:31:12.925800 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.925944 kubelet[2683]: E0310 01:31:12.925828 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.926584 kubelet[2683]: E0310 01:31:12.926566 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.926696 kubelet[2683]: W0310 01:31:12.926672 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.926820 kubelet[2683]: E0310 01:31:12.926802 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.928142 kubelet[2683]: E0310 01:31:12.928069 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.928142 kubelet[2683]: W0310 01:31:12.928087 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.928142 kubelet[2683]: E0310 01:31:12.928106 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.946291 kubelet[2683]: E0310 01:31:12.946256 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 10 01:31:12.946291 kubelet[2683]: W0310 01:31:12.946280 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 10 01:31:12.946291 kubelet[2683]: E0310 01:31:12.946301 2683 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 10 01:31:12.951103 containerd[1507]: time="2026-03-10T01:31:12.950865970Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:12.953746 containerd[1507]: time="2026-03-10T01:31:12.952633350Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:12.953746 containerd[1507]: time="2026-03-10T01:31:12.952692060Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:12.953746 containerd[1507]: time="2026-03-10T01:31:12.952939928Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:12.980455 systemd[1]: Started cri-containerd-b0112815a966663c4b60e7c7680c757cc6dde1e3a96aed8d2d1be064afc4aa5f.scope - libcontainer container b0112815a966663c4b60e7c7680c757cc6dde1e3a96aed8d2d1be064afc4aa5f. Mar 10 01:31:12.999423 systemd[1]: Started cri-containerd-37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc.scope - libcontainer container 37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc. Mar 10 01:31:13.060981 containerd[1507]: time="2026-03-10T01:31:13.060803779Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-7sq4w,Uid:bd178a7f-8866-4467-8b65-479d48e91fb8,Namespace:calico-system,Attempt:0,} returns sandbox id \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\"" Mar 10 01:31:13.084180 containerd[1507]: time="2026-03-10T01:31:13.084138456Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\"" Mar 10 01:31:13.102645 containerd[1507]: time="2026-03-10T01:31:13.102378024Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6787995c7-7f9vg,Uid:51c5114e-9642-4a30-8ebb-f32b5f920061,Namespace:calico-system,Attempt:0,} returns sandbox id \"b0112815a966663c4b60e7c7680c757cc6dde1e3a96aed8d2d1be064afc4aa5f\"" Mar 10 01:31:14.225262 kubelet[2683]: E0310 01:31:14.225143 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:14.713439 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3069519838.mount: Deactivated successfully. Mar 10 01:31:14.857257 containerd[1507]: time="2026-03-10T01:31:14.855560637Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:14.857257 containerd[1507]: time="2026-03-10T01:31:14.856766265Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4: active requests=0, bytes read=6186433" Mar 10 01:31:14.857257 containerd[1507]: time="2026-03-10T01:31:14.856794810Z" level=info msg="ImageCreate event name:\"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:14.860806 containerd[1507]: time="2026-03-10T01:31:14.860765667Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:14.862106 containerd[1507]: time="2026-03-10T01:31:14.862064751Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" with image id \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:5fa3492ac4dfef9cc34fe70a51289118e1f715a89133ea730eef81ad789dadbc\", size \"6186255\" in 1.777765605s" Mar 10 01:31:14.862165 containerd[1507]: time="2026-03-10T01:31:14.862107654Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.31.4\" returns image reference \"sha256:a6ea0cf732d820506ae9f1d7e7433a14009026b894fbbb8f346b9a5f5335c47e\"" Mar 10 01:31:14.865030 containerd[1507]: time="2026-03-10T01:31:14.864969348Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\"" Mar 10 01:31:14.867500 containerd[1507]: time="2026-03-10T01:31:14.867469945Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 10 01:31:14.885204 containerd[1507]: time="2026-03-10T01:31:14.885168133Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998\"" Mar 10 01:31:14.886171 containerd[1507]: time="2026-03-10T01:31:14.886138737Z" level=info msg="StartContainer for \"a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998\"" Mar 10 01:31:14.940041 systemd[1]: run-containerd-runc-k8s.io-a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998-runc.igNGaw.mount: Deactivated successfully. Mar 10 01:31:14.953464 systemd[1]: Started cri-containerd-a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998.scope - libcontainer container a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998. Mar 10 01:31:15.006153 containerd[1507]: time="2026-03-10T01:31:15.005863303Z" level=info msg="StartContainer for \"a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998\" returns successfully" Mar 10 01:31:15.024500 systemd[1]: cri-containerd-a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998.scope: Deactivated successfully. Mar 10 01:31:15.130789 containerd[1507]: time="2026-03-10T01:31:15.105181175Z" level=info msg="shim disconnected" id=a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998 namespace=k8s.io Mar 10 01:31:15.131680 containerd[1507]: time="2026-03-10T01:31:15.131390556Z" level=warning msg="cleaning up after shim disconnected" id=a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998 namespace=k8s.io Mar 10 01:31:15.131680 containerd[1507]: time="2026-03-10T01:31:15.131423974Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 10 01:31:15.714350 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a9219078bd11cbf695d348896c19b568a16af122718769c275a55a9a71d94998-rootfs.mount: Deactivated successfully. Mar 10 01:31:16.226267 kubelet[2683]: E0310 01:31:16.224651 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:18.111914 containerd[1507]: time="2026-03-10T01:31:18.110433351Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:18.114471 containerd[1507]: time="2026-03-10T01:31:18.114415227Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.31.4: active requests=0, bytes read=34551413" Mar 10 01:31:18.115714 containerd[1507]: time="2026-03-10T01:31:18.115661837Z" level=info msg="ImageCreate event name:\"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:18.120528 containerd[1507]: time="2026-03-10T01:31:18.120499176Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:18.121788 containerd[1507]: time="2026-03-10T01:31:18.121737708Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.31.4\" with image id \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\", repo tag \"ghcr.io/flatcar/calico/typha:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d9396cfcd63dfcf72a65903042e473bb0bafc0cceb56bd71cd84078498a87130\", size \"36107450\" in 3.255844701s" Mar 10 01:31:18.121877 containerd[1507]: time="2026-03-10T01:31:18.121810299Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.31.4\" returns image reference \"sha256:46766605472b59b9c16342b2cc74da11f598baa9ba6d1e8b07b3f8ab4f29c55b\"" Mar 10 01:31:18.125134 containerd[1507]: time="2026-03-10T01:31:18.124582501Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\"" Mar 10 01:31:18.149255 containerd[1507]: time="2026-03-10T01:31:18.149169638Z" level=info msg="CreateContainer within sandbox \"b0112815a966663c4b60e7c7680c757cc6dde1e3a96aed8d2d1be064afc4aa5f\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 10 01:31:18.168601 containerd[1507]: time="2026-03-10T01:31:18.168359674Z" level=info msg="CreateContainer within sandbox \"b0112815a966663c4b60e7c7680c757cc6dde1e3a96aed8d2d1be064afc4aa5f\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"2df4b82bfa9f4a1ccd49c612b0089ba105beffac8701cdc01573a3bf5412eef2\"" Mar 10 01:31:18.170178 containerd[1507]: time="2026-03-10T01:31:18.169285866Z" level=info msg="StartContainer for \"2df4b82bfa9f4a1ccd49c612b0089ba105beffac8701cdc01573a3bf5412eef2\"" Mar 10 01:31:18.227361 kubelet[2683]: E0310 01:31:18.225788 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:18.250463 systemd[1]: Started cri-containerd-2df4b82bfa9f4a1ccd49c612b0089ba105beffac8701cdc01573a3bf5412eef2.scope - libcontainer container 2df4b82bfa9f4a1ccd49c612b0089ba105beffac8701cdc01573a3bf5412eef2. Mar 10 01:31:18.353046 containerd[1507]: time="2026-03-10T01:31:18.352816929Z" level=info msg="StartContainer for \"2df4b82bfa9f4a1ccd49c612b0089ba105beffac8701cdc01573a3bf5412eef2\" returns successfully" Mar 10 01:31:18.421627 kubelet[2683]: I0310 01:31:18.420357 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6787995c7-7f9vg" podStartSLOduration=1.401851302 podStartE2EDuration="6.420319003s" podCreationTimestamp="2026-03-10 01:31:12 +0000 UTC" firstStartedPulling="2026-03-10 01:31:13.105125827 +0000 UTC m=+23.212608930" lastFinishedPulling="2026-03-10 01:31:18.123593517 +0000 UTC m=+28.231076631" observedRunningTime="2026-03-10 01:31:18.419461508 +0000 UTC m=+28.526944634" watchObservedRunningTime="2026-03-10 01:31:18.420319003 +0000 UTC m=+28.527802110" Mar 10 01:31:20.227192 kubelet[2683]: E0310 01:31:20.226159 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:22.226263 kubelet[2683]: E0310 01:31:22.225470 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:24.226194 kubelet[2683]: E0310 01:31:24.226056 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:26.225186 kubelet[2683]: E0310 01:31:26.224279 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:28.227254 kubelet[2683]: E0310 01:31:28.226887 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:29.485050 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4127876815.mount: Deactivated successfully. Mar 10 01:31:29.554241 containerd[1507]: time="2026-03-10T01:31:29.548805105Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.31.4: active requests=0, bytes read=159838564" Mar 10 01:31:29.555454 containerd[1507]: time="2026-03-10T01:31:29.547507148Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:29.559312 containerd[1507]: time="2026-03-10T01:31:29.559279048Z" level=info msg="ImageCreate event name:\"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:29.561481 containerd[1507]: time="2026-03-10T01:31:29.561436221Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.31.4\" with image id \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\", repo tag \"ghcr.io/flatcar/calico/node:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\", size \"159838426\" in 11.436807668s" Mar 10 01:31:29.561583 containerd[1507]: time="2026-03-10T01:31:29.561506281Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.31.4\" returns image reference \"sha256:e6536b93706eda782f82ebadcac3559cb61801d09f982cc0533a134e6a8e1acf\"" Mar 10 01:31:29.562430 containerd[1507]: time="2026-03-10T01:31:29.562395577Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:22b9d32dc7480c96272121d5682d53424c6e58653c60fa869b61a1758a11d77f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:29.571386 containerd[1507]: time="2026-03-10T01:31:29.569394266Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for container &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,}" Mar 10 01:31:29.597931 containerd[1507]: time="2026-03-10T01:31:29.597880946Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for &ContainerMetadata{Name:ebpf-bootstrap,Attempt:0,} returns container id \"e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979\"" Mar 10 01:31:29.600183 containerd[1507]: time="2026-03-10T01:31:29.600127637Z" level=info msg="StartContainer for \"e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979\"" Mar 10 01:31:29.695742 systemd[1]: Started cri-containerd-e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979.scope - libcontainer container e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979. Mar 10 01:31:29.765308 containerd[1507]: time="2026-03-10T01:31:29.763303829Z" level=info msg="StartContainer for \"e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979\" returns successfully" Mar 10 01:31:30.009735 systemd[1]: cri-containerd-e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979.scope: Deactivated successfully. Mar 10 01:31:30.075121 containerd[1507]: time="2026-03-10T01:31:30.072882149Z" level=info msg="shim disconnected" id=e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979 namespace=k8s.io Mar 10 01:31:30.075121 containerd[1507]: time="2026-03-10T01:31:30.075006431Z" level=warning msg="cleaning up after shim disconnected" id=e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979 namespace=k8s.io Mar 10 01:31:30.075121 containerd[1507]: time="2026-03-10T01:31:30.075033218Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 10 01:31:30.101932 containerd[1507]: time="2026-03-10T01:31:30.101834762Z" level=warning msg="cleanup warnings time=\"2026-03-10T01:31:30Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Mar 10 01:31:30.228972 kubelet[2683]: E0310 01:31:30.228500 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:30.436444 containerd[1507]: time="2026-03-10T01:31:30.436344302Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\"" Mar 10 01:31:30.479854 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e6cf238f0ef0ef2a2e46b5a487421a8acf1ce427959fd0dd6d86f7fe4b73e979-rootfs.mount: Deactivated successfully. Mar 10 01:31:32.227573 kubelet[2683]: E0310 01:31:32.227414 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:34.234583 kubelet[2683]: E0310 01:31:34.234440 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:35.360266 containerd[1507]: time="2026-03-10T01:31:35.358507059Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:35.360266 containerd[1507]: time="2026-03-10T01:31:35.359927698Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.31.4: active requests=0, bytes read=70611671" Mar 10 01:31:35.361344 containerd[1507]: time="2026-03-10T01:31:35.360209851Z" level=info msg="ImageCreate event name:\"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:35.367160 containerd[1507]: time="2026-03-10T01:31:35.367125715Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:35.368822 containerd[1507]: time="2026-03-10T01:31:35.368429230Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.31.4\" with image id \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\", repo tag \"ghcr.io/flatcar/calico/cni:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:f1c5d9a6df01061c5faec4c4b59fb9ba69f8f5164b51e01ea8daa8e373111a04\", size \"72167716\" in 4.932014081s" Mar 10 01:31:35.368822 containerd[1507]: time="2026-03-10T01:31:35.368483507Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.31.4\" returns image reference \"sha256:c433a27dd94ce9242338eece49f11629412dd42552fed314746fcf16ea958b2b\"" Mar 10 01:31:35.410249 containerd[1507]: time="2026-03-10T01:31:35.409841393Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 10 01:31:35.436745 containerd[1507]: time="2026-03-10T01:31:35.436696603Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e\"" Mar 10 01:31:35.438783 containerd[1507]: time="2026-03-10T01:31:35.438747880Z" level=info msg="StartContainer for \"3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e\"" Mar 10 01:31:35.512398 systemd[1]: Started cri-containerd-3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e.scope - libcontainer container 3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e. Mar 10 01:31:35.578477 containerd[1507]: time="2026-03-10T01:31:35.577735783Z" level=info msg="StartContainer for \"3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e\" returns successfully" Mar 10 01:31:36.225839 kubelet[2683]: E0310 01:31:36.225764 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:36.890960 systemd[1]: cri-containerd-3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e.scope: Deactivated successfully. Mar 10 01:31:36.939983 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e-rootfs.mount: Deactivated successfully. Mar 10 01:31:36.985576 containerd[1507]: time="2026-03-10T01:31:36.985457893Z" level=info msg="shim disconnected" id=3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e namespace=k8s.io Mar 10 01:31:36.986510 containerd[1507]: time="2026-03-10T01:31:36.985576277Z" level=warning msg="cleaning up after shim disconnected" id=3e38bc578551c84f25bd05ddca5384a1854c9faf4d39a9fc87a51393c85fa92e namespace=k8s.io Mar 10 01:31:36.986510 containerd[1507]: time="2026-03-10T01:31:36.985600986Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 10 01:31:36.991458 kubelet[2683]: I0310 01:31:36.988750 2683 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Mar 10 01:31:37.129836 systemd[1]: Created slice kubepods-burstable-podea8ee6c7_d14f_40b8_b33b_78317f32f0c3.slice - libcontainer container kubepods-burstable-podea8ee6c7_d14f_40b8_b33b_78317f32f0c3.slice. Mar 10 01:31:37.152576 systemd[1]: Created slice kubepods-besteffort-podc035353a_3dfd_42e9_a291_75ad204d0496.slice - libcontainer container kubepods-besteffort-podc035353a_3dfd_42e9_a291_75ad204d0496.slice. Mar 10 01:31:37.167874 systemd[1]: Created slice kubepods-besteffort-pod2c9450c8_37a4_4965_96d3_67fee0cc59e8.slice - libcontainer container kubepods-besteffort-pod2c9450c8_37a4_4965_96d3_67fee0cc59e8.slice. Mar 10 01:31:37.185351 systemd[1]: Created slice kubepods-besteffort-pod7724f152_38e8_4ee2_b7cd_2755c7f51836.slice - libcontainer container kubepods-besteffort-pod7724f152_38e8_4ee2_b7cd_2755c7f51836.slice. Mar 10 01:31:37.202173 systemd[1]: Created slice kubepods-burstable-pod8c562e47_a863_4e0d_8c19_218349a688f3.slice - libcontainer container kubepods-burstable-pod8c562e47_a863_4e0d_8c19_218349a688f3.slice. Mar 10 01:31:37.223223 systemd[1]: Created slice kubepods-besteffort-pod4c86706d_5198_4966_aa4f_480259ec456a.slice - libcontainer container kubepods-besteffort-pod4c86706d_5198_4966_aa4f_480259ec456a.slice. Mar 10 01:31:37.240918 kubelet[2683]: I0310 01:31:37.238776 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zshxk\" (UniqueName: \"kubernetes.io/projected/2c9450c8-37a4-4965-96d3-67fee0cc59e8-kube-api-access-zshxk\") pod \"calico-kube-controllers-77f6556c5c-68b2f\" (UID: \"2c9450c8-37a4-4965-96d3-67fee0cc59e8\") " pod="calico-system/calico-kube-controllers-77f6556c5c-68b2f" Mar 10 01:31:37.240918 kubelet[2683]: I0310 01:31:37.238845 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c86706d-5198-4966-aa4f-480259ec456a-goldmane-ca-bundle\") pod \"goldmane-cccfbd5cf-sxw9t\" (UID: \"4c86706d-5198-4966-aa4f-480259ec456a\") " pod="calico-system/goldmane-cccfbd5cf-sxw9t" Mar 10 01:31:37.240918 kubelet[2683]: I0310 01:31:37.238886 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/4c86706d-5198-4966-aa4f-480259ec456a-goldmane-key-pair\") pod \"goldmane-cccfbd5cf-sxw9t\" (UID: \"4c86706d-5198-4966-aa4f-480259ec456a\") " pod="calico-system/goldmane-cccfbd5cf-sxw9t" Mar 10 01:31:37.240918 kubelet[2683]: I0310 01:31:37.238912 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99mfh\" (UniqueName: \"kubernetes.io/projected/4c86706d-5198-4966-aa4f-480259ec456a-kube-api-access-99mfh\") pod \"goldmane-cccfbd5cf-sxw9t\" (UID: \"4c86706d-5198-4966-aa4f-480259ec456a\") " pod="calico-system/goldmane-cccfbd5cf-sxw9t" Mar 10 01:31:37.240918 kubelet[2683]: I0310 01:31:37.238957 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r69q4\" (UniqueName: \"kubernetes.io/projected/8c562e47-a863-4e0d-8c19-218349a688f3-kube-api-access-r69q4\") pod \"coredns-66bc5c9577-jjvj8\" (UID: \"8c562e47-a863-4e0d-8c19-218349a688f3\") " pod="kube-system/coredns-66bc5c9577-jjvj8" Mar 10 01:31:37.240491 systemd[1]: Created slice kubepods-besteffort-podfa9659b6_52d9_4a4d_91d1_6f294264f5ee.slice - libcontainer container kubepods-besteffort-podfa9659b6_52d9_4a4d_91d1_6f294264f5ee.slice. Mar 10 01:31:37.242492 kubelet[2683]: I0310 01:31:37.238994 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c86706d-5198-4966-aa4f-480259ec456a-config\") pod \"goldmane-cccfbd5cf-sxw9t\" (UID: \"4c86706d-5198-4966-aa4f-480259ec456a\") " pod="calico-system/goldmane-cccfbd5cf-sxw9t" Mar 10 01:31:37.242492 kubelet[2683]: I0310 01:31:37.239033 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea8ee6c7-d14f-40b8-b33b-78317f32f0c3-config-volume\") pod \"coredns-66bc5c9577-2d4ds\" (UID: \"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3\") " pod="kube-system/coredns-66bc5c9577-2d4ds" Mar 10 01:31:37.242492 kubelet[2683]: I0310 01:31:37.239060 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c9450c8-37a4-4965-96d3-67fee0cc59e8-tigera-ca-bundle\") pod \"calico-kube-controllers-77f6556c5c-68b2f\" (UID: \"2c9450c8-37a4-4965-96d3-67fee0cc59e8\") " pod="calico-system/calico-kube-controllers-77f6556c5c-68b2f" Mar 10 01:31:37.242492 kubelet[2683]: I0310 01:31:37.239086 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c035353a-3dfd-42e9-a291-75ad204d0496-calico-apiserver-certs\") pod \"calico-apiserver-794866df9b-424tn\" (UID: \"c035353a-3dfd-42e9-a291-75ad204d0496\") " pod="calico-system/calico-apiserver-794866df9b-424tn" Mar 10 01:31:37.242492 kubelet[2683]: I0310 01:31:37.239137 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-backend-key-pair\") pod \"whisker-7475b8444d-2s95h\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " pod="calico-system/whisker-7475b8444d-2s95h" Mar 10 01:31:37.242739 kubelet[2683]: I0310 01:31:37.239166 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbgwb\" (UniqueName: \"kubernetes.io/projected/ea8ee6c7-d14f-40b8-b33b-78317f32f0c3-kube-api-access-nbgwb\") pod \"coredns-66bc5c9577-2d4ds\" (UID: \"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3\") " pod="kube-system/coredns-66bc5c9577-2d4ds" Mar 10 01:31:37.242739 kubelet[2683]: I0310 01:31:37.239204 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ng5h\" (UniqueName: \"kubernetes.io/projected/c035353a-3dfd-42e9-a291-75ad204d0496-kube-api-access-2ng5h\") pod \"calico-apiserver-794866df9b-424tn\" (UID: \"c035353a-3dfd-42e9-a291-75ad204d0496\") " pod="calico-system/calico-apiserver-794866df9b-424tn" Mar 10 01:31:37.242739 kubelet[2683]: I0310 01:31:37.239262 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-nginx-config\") pod \"whisker-7475b8444d-2s95h\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " pod="calico-system/whisker-7475b8444d-2s95h" Mar 10 01:31:37.242739 kubelet[2683]: I0310 01:31:37.239296 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-ca-bundle\") pod \"whisker-7475b8444d-2s95h\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " pod="calico-system/whisker-7475b8444d-2s95h" Mar 10 01:31:37.242739 kubelet[2683]: I0310 01:31:37.239328 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4hft\" (UniqueName: \"kubernetes.io/projected/7724f152-38e8-4ee2-b7cd-2755c7f51836-kube-api-access-n4hft\") pod \"whisker-7475b8444d-2s95h\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " pod="calico-system/whisker-7475b8444d-2s95h" Mar 10 01:31:37.242976 kubelet[2683]: I0310 01:31:37.239393 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/fa9659b6-52d9-4a4d-91d1-6f294264f5ee-calico-apiserver-certs\") pod \"calico-apiserver-794866df9b-h52bq\" (UID: \"fa9659b6-52d9-4a4d-91d1-6f294264f5ee\") " pod="calico-system/calico-apiserver-794866df9b-h52bq" Mar 10 01:31:37.242976 kubelet[2683]: I0310 01:31:37.239425 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpbf2\" (UniqueName: \"kubernetes.io/projected/fa9659b6-52d9-4a4d-91d1-6f294264f5ee-kube-api-access-gpbf2\") pod \"calico-apiserver-794866df9b-h52bq\" (UID: \"fa9659b6-52d9-4a4d-91d1-6f294264f5ee\") " pod="calico-system/calico-apiserver-794866df9b-h52bq" Mar 10 01:31:37.242976 kubelet[2683]: I0310 01:31:37.239486 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c562e47-a863-4e0d-8c19-218349a688f3-config-volume\") pod \"coredns-66bc5c9577-jjvj8\" (UID: \"8c562e47-a863-4e0d-8c19-218349a688f3\") " pod="kube-system/coredns-66bc5c9577-jjvj8" Mar 10 01:31:37.460753 containerd[1507]: time="2026-03-10T01:31:37.460538664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2d4ds,Uid:ea8ee6c7-d14f-40b8-b33b-78317f32f0c3,Namespace:kube-system,Attempt:0,}" Mar 10 01:31:37.463900 containerd[1507]: time="2026-03-10T01:31:37.463802133Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-424tn,Uid:c035353a-3dfd-42e9-a291-75ad204d0496,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:37.498688 containerd[1507]: time="2026-03-10T01:31:37.498463804Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-77f6556c5c-68b2f,Uid:2c9450c8-37a4-4965-96d3-67fee0cc59e8,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:37.502915 containerd[1507]: time="2026-03-10T01:31:37.502344481Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7475b8444d-2s95h,Uid:7724f152-38e8-4ee2-b7cd-2755c7f51836,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:37.537765 containerd[1507]: time="2026-03-10T01:31:37.537164900Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-jjvj8,Uid:8c562e47-a863-4e0d-8c19-218349a688f3,Namespace:kube-system,Attempt:0,}" Mar 10 01:31:37.539422 containerd[1507]: time="2026-03-10T01:31:37.539390060Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 10 01:31:37.548242 containerd[1507]: time="2026-03-10T01:31:37.546811954Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-cccfbd5cf-sxw9t,Uid:4c86706d-5198-4966-aa4f-480259ec456a,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:37.558131 containerd[1507]: time="2026-03-10T01:31:37.557929568Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-h52bq,Uid:fa9659b6-52d9-4a4d-91d1-6f294264f5ee,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:37.697443 containerd[1507]: time="2026-03-10T01:31:37.696719859Z" level=info msg="CreateContainer within sandbox \"37ccd8546d560bee0954c198099c679289a4cea5ea6ac76d947fe6fac36864dc\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"828b65ed9f2fa67d61efbbe6be7fe5256bc16a55d277a6b804b27a98526d1147\"" Mar 10 01:31:37.722018 containerd[1507]: time="2026-03-10T01:31:37.721434196Z" level=info msg="StartContainer for \"828b65ed9f2fa67d61efbbe6be7fe5256bc16a55d277a6b804b27a98526d1147\"" Mar 10 01:31:37.929473 systemd[1]: Started cri-containerd-828b65ed9f2fa67d61efbbe6be7fe5256bc16a55d277a6b804b27a98526d1147.scope - libcontainer container 828b65ed9f2fa67d61efbbe6be7fe5256bc16a55d277a6b804b27a98526d1147. Mar 10 01:31:38.127441 containerd[1507]: time="2026-03-10T01:31:38.127316068Z" level=error msg="Failed to destroy network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.133907 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8-shm.mount: Deactivated successfully. Mar 10 01:31:38.149164 containerd[1507]: time="2026-03-10T01:31:38.149067557Z" level=error msg="Failed to destroy network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.151173 containerd[1507]: time="2026-03-10T01:31:38.150269015Z" level=error msg="encountered an error cleaning up failed sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.151173 containerd[1507]: time="2026-03-10T01:31:38.150389677Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7475b8444d-2s95h,Uid:7724f152-38e8-4ee2-b7cd-2755c7f51836,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.158025 containerd[1507]: time="2026-03-10T01:31:38.154460027Z" level=error msg="Failed to destroy network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.155802 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1-shm.mount: Deactivated successfully. Mar 10 01:31:38.167745 containerd[1507]: time="2026-03-10T01:31:38.167160091Z" level=error msg="encountered an error cleaning up failed sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.168694 containerd[1507]: time="2026-03-10T01:31:38.168633879Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2d4ds,Uid:ea8ee6c7-d14f-40b8-b33b-78317f32f0c3,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.169568 containerd[1507]: time="2026-03-10T01:31:38.169295153Z" level=error msg="encountered an error cleaning up failed sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.170008 containerd[1507]: time="2026-03-10T01:31:38.169942263Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-cccfbd5cf-sxw9t,Uid:4c86706d-5198-4966-aa4f-480259ec456a,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.171735 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be-shm.mount: Deactivated successfully. Mar 10 01:31:38.182392 containerd[1507]: time="2026-03-10T01:31:38.182336820Z" level=info msg="StartContainer for \"828b65ed9f2fa67d61efbbe6be7fe5256bc16a55d277a6b804b27a98526d1147\" returns successfully" Mar 10 01:31:38.221296 kubelet[2683]: E0310 01:31:38.220667 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.221746 kubelet[2683]: E0310 01:31:38.221562 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.223746 kubelet[2683]: E0310 01:31:38.222187 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7475b8444d-2s95h" Mar 10 01:31:38.223746 kubelet[2683]: E0310 01:31:38.223383 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7475b8444d-2s95h" Mar 10 01:31:38.223746 kubelet[2683]: E0310 01:31:38.223390 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.223746 kubelet[2683]: E0310 01:31:38.223438 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-cccfbd5cf-sxw9t" Mar 10 01:31:38.223986 kubelet[2683]: E0310 01:31:38.223473 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-cccfbd5cf-sxw9t" Mar 10 01:31:38.223986 kubelet[2683]: E0310 01:31:38.223501 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7475b8444d-2s95h_calico-system(7724f152-38e8-4ee2-b7cd-2755c7f51836)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7475b8444d-2s95h_calico-system(7724f152-38e8-4ee2-b7cd-2755c7f51836)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7475b8444d-2s95h" podUID="7724f152-38e8-4ee2-b7cd-2755c7f51836" Mar 10 01:31:38.223986 kubelet[2683]: E0310 01:31:38.223540 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-cccfbd5cf-sxw9t_calico-system(4c86706d-5198-4966-aa4f-480259ec456a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-cccfbd5cf-sxw9t_calico-system(4c86706d-5198-4966-aa4f-480259ec456a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-cccfbd5cf-sxw9t" podUID="4c86706d-5198-4966-aa4f-480259ec456a" Mar 10 01:31:38.228202 kubelet[2683]: E0310 01:31:38.222258 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-2d4ds" Mar 10 01:31:38.229333 kubelet[2683]: E0310 01:31:38.229281 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-2d4ds" Mar 10 01:31:38.230433 kubelet[2683]: E0310 01:31:38.230388 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-2d4ds_kube-system(ea8ee6c7-d14f-40b8-b33b-78317f32f0c3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-2d4ds_kube-system(ea8ee6c7-d14f-40b8-b33b-78317f32f0c3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-2d4ds" podUID="ea8ee6c7-d14f-40b8-b33b-78317f32f0c3" Mar 10 01:31:38.233187 containerd[1507]: time="2026-03-10T01:31:38.233133321Z" level=error msg="Failed to destroy network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.244039 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1-shm.mount: Deactivated successfully. Mar 10 01:31:38.247476 containerd[1507]: time="2026-03-10T01:31:38.247109269Z" level=error msg="encountered an error cleaning up failed sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.251062 containerd[1507]: time="2026-03-10T01:31:38.247216050Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-77f6556c5c-68b2f,Uid:2c9450c8-37a4-4965-96d3-67fee0cc59e8,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.253296 kubelet[2683]: E0310 01:31:38.252197 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.253296 kubelet[2683]: E0310 01:31:38.253212 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-77f6556c5c-68b2f" Mar 10 01:31:38.256719 kubelet[2683]: E0310 01:31:38.254212 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-77f6556c5c-68b2f" Mar 10 01:31:38.256719 kubelet[2683]: E0310 01:31:38.254369 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-77f6556c5c-68b2f_calico-system(2c9450c8-37a4-4965-96d3-67fee0cc59e8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-77f6556c5c-68b2f_calico-system(2c9450c8-37a4-4965-96d3-67fee0cc59e8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-77f6556c5c-68b2f" podUID="2c9450c8-37a4-4965-96d3-67fee0cc59e8" Mar 10 01:31:38.261508 systemd[1]: Created slice kubepods-besteffort-pod71ea22a7_e0cb_4bd4_871f_854f4db6a991.slice - libcontainer container kubepods-besteffort-pod71ea22a7_e0cb_4bd4_871f_854f4db6a991.slice. Mar 10 01:31:38.274604 containerd[1507]: time="2026-03-10T01:31:38.274551503Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fhcl2,Uid:71ea22a7-e0cb-4bd4-871f-854f4db6a991,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:38.301159 containerd[1507]: time="2026-03-10T01:31:38.301098366Z" level=error msg="Failed to destroy network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.301963 containerd[1507]: time="2026-03-10T01:31:38.301927269Z" level=error msg="encountered an error cleaning up failed sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.302135 containerd[1507]: time="2026-03-10T01:31:38.302096308Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-h52bq,Uid:fa9659b6-52d9-4a4d-91d1-6f294264f5ee,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.304332 kubelet[2683]: E0310 01:31:38.303479 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.304332 kubelet[2683]: E0310 01:31:38.303573 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-794866df9b-h52bq" Mar 10 01:31:38.304332 kubelet[2683]: E0310 01:31:38.303615 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-794866df9b-h52bq" Mar 10 01:31:38.304528 kubelet[2683]: E0310 01:31:38.303729 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-794866df9b-h52bq_calico-system(fa9659b6-52d9-4a4d-91d1-6f294264f5ee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-794866df9b-h52bq_calico-system(fa9659b6-52d9-4a4d-91d1-6f294264f5ee)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-794866df9b-h52bq" podUID="fa9659b6-52d9-4a4d-91d1-6f294264f5ee" Mar 10 01:31:38.328807 containerd[1507]: time="2026-03-10T01:31:38.328442222Z" level=error msg="Failed to destroy network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.331334 containerd[1507]: time="2026-03-10T01:31:38.331063276Z" level=error msg="encountered an error cleaning up failed sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.331847 containerd[1507]: time="2026-03-10T01:31:38.331616831Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-424tn,Uid:c035353a-3dfd-42e9-a291-75ad204d0496,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.332474 kubelet[2683]: E0310 01:31:38.332357 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.334352 kubelet[2683]: E0310 01:31:38.332683 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-794866df9b-424tn" Mar 10 01:31:38.334352 kubelet[2683]: E0310 01:31:38.332722 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-apiserver-794866df9b-424tn" Mar 10 01:31:38.334352 kubelet[2683]: E0310 01:31:38.332805 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-794866df9b-424tn_calico-system(c035353a-3dfd-42e9-a291-75ad204d0496)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-794866df9b-424tn_calico-system(c035353a-3dfd-42e9-a291-75ad204d0496)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-794866df9b-424tn" podUID="c035353a-3dfd-42e9-a291-75ad204d0496" Mar 10 01:31:38.341187 containerd[1507]: time="2026-03-10T01:31:38.341061221Z" level=error msg="Failed to destroy network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.341676 containerd[1507]: time="2026-03-10T01:31:38.341617443Z" level=error msg="encountered an error cleaning up failed sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.341762 containerd[1507]: time="2026-03-10T01:31:38.341695239Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-jjvj8,Uid:8c562e47-a863-4e0d-8c19-218349a688f3,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.342384 kubelet[2683]: E0310 01:31:38.342322 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.342764 kubelet[2683]: E0310 01:31:38.342562 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-jjvj8" Mar 10 01:31:38.342764 kubelet[2683]: E0310 01:31:38.342614 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-jjvj8" Mar 10 01:31:38.342963 kubelet[2683]: E0310 01:31:38.342894 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-jjvj8_kube-system(8c562e47-a863-4e0d-8c19-218349a688f3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-jjvj8_kube-system(8c562e47-a863-4e0d-8c19-218349a688f3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-jjvj8" podUID="8c562e47-a863-4e0d-8c19-218349a688f3" Mar 10 01:31:38.430279 containerd[1507]: time="2026-03-10T01:31:38.430046482Z" level=error msg="Failed to destroy network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.432578 containerd[1507]: time="2026-03-10T01:31:38.432503023Z" level=error msg="encountered an error cleaning up failed sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.432677 containerd[1507]: time="2026-03-10T01:31:38.432602529Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fhcl2,Uid:71ea22a7-e0cb-4bd4-871f-854f4db6a991,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.433476 kubelet[2683]: E0310 01:31:38.432967 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.433476 kubelet[2683]: E0310 01:31:38.433073 2683 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-fhcl2" Mar 10 01:31:38.433476 kubelet[2683]: E0310 01:31:38.433113 2683 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-fhcl2" Mar 10 01:31:38.433676 kubelet[2683]: E0310 01:31:38.433211 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-fhcl2_calico-system(71ea22a7-e0cb-4bd4-871f-854f4db6a991)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-fhcl2_calico-system(71ea22a7-e0cb-4bd4-871f-854f4db6a991)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:38.481188 kubelet[2683]: I0310 01:31:38.481108 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:31:38.487868 kubelet[2683]: I0310 01:31:38.487673 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:31:38.502765 kubelet[2683]: I0310 01:31:38.501832 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:31:38.526168 kubelet[2683]: I0310 01:31:38.525232 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:31:38.532903 kubelet[2683]: I0310 01:31:38.532337 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:31:38.536732 kubelet[2683]: I0310 01:31:38.536029 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:38.542449 containerd[1507]: time="2026-03-10T01:31:38.542385095Z" level=info msg="StopPodSandbox for \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\"" Mar 10 01:31:38.544406 containerd[1507]: time="2026-03-10T01:31:38.544333276Z" level=info msg="StopPodSandbox for \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\"" Mar 10 01:31:38.544714 containerd[1507]: time="2026-03-10T01:31:38.544683409Z" level=info msg="StopPodSandbox for \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\"" Mar 10 01:31:38.544835 containerd[1507]: time="2026-03-10T01:31:38.544796926Z" level=info msg="Ensure that sandbox 42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1 in task-service has been cleanup successfully" Mar 10 01:31:38.547305 containerd[1507]: time="2026-03-10T01:31:38.547263777Z" level=info msg="Ensure that sandbox c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8 in task-service has been cleanup successfully" Mar 10 01:31:38.554843 containerd[1507]: time="2026-03-10T01:31:38.554797061Z" level=info msg="StopPodSandbox for \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\"" Mar 10 01:31:38.555607 containerd[1507]: time="2026-03-10T01:31:38.555265050Z" level=info msg="Ensure that sandbox 1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3 in task-service has been cleanup successfully" Mar 10 01:31:38.557704 kubelet[2683]: I0310 01:31:38.557002 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:31:38.558724 containerd[1507]: time="2026-03-10T01:31:38.544359898Z" level=info msg="Ensure that sandbox e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0 in task-service has been cleanup successfully" Mar 10 01:31:38.559636 containerd[1507]: time="2026-03-10T01:31:38.559598586Z" level=info msg="StopPodSandbox for \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\"" Mar 10 01:31:38.576683 containerd[1507]: time="2026-03-10T01:31:38.576608545Z" level=info msg="StopPodSandbox for \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\"" Mar 10 01:31:38.576916 containerd[1507]: time="2026-03-10T01:31:38.576887287Z" level=info msg="Ensure that sandbox e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1 in task-service has been cleanup successfully" Mar 10 01:31:38.592478 containerd[1507]: time="2026-03-10T01:31:38.591186962Z" level=info msg="Ensure that sandbox 50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394 in task-service has been cleanup successfully" Mar 10 01:31:38.598462 containerd[1507]: time="2026-03-10T01:31:38.596639986Z" level=info msg="StopPodSandbox for \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\"" Mar 10 01:31:38.598462 containerd[1507]: time="2026-03-10T01:31:38.597194010Z" level=info msg="Ensure that sandbox d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be in task-service has been cleanup successfully" Mar 10 01:31:38.684398 kubelet[2683]: I0310 01:31:38.682114 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:31:38.685164 containerd[1507]: time="2026-03-10T01:31:38.684884100Z" level=info msg="StopPodSandbox for \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\"" Mar 10 01:31:38.686719 containerd[1507]: time="2026-03-10T01:31:38.686626239Z" level=info msg="Ensure that sandbox 21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f in task-service has been cleanup successfully" Mar 10 01:31:38.708709 kubelet[2683]: I0310 01:31:38.708534 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-7sq4w" podStartSLOduration=4.383150427 podStartE2EDuration="26.693130389s" podCreationTimestamp="2026-03-10 01:31:12 +0000 UTC" firstStartedPulling="2026-03-10 01:31:13.064786554 +0000 UTC m=+23.172269658" lastFinishedPulling="2026-03-10 01:31:35.374766519 +0000 UTC m=+45.482249620" observedRunningTime="2026-03-10 01:31:38.681262028 +0000 UTC m=+48.788745173" watchObservedRunningTime="2026-03-10 01:31:38.693130389 +0000 UTC m=+48.800613503" Mar 10 01:31:38.795803 containerd[1507]: time="2026-03-10T01:31:38.795639440Z" level=error msg="StopPodSandbox for \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\" failed" error="failed to destroy network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.797311 kubelet[2683]: E0310 01:31:38.796999 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:31:38.800019 kubelet[2683]: E0310 01:31:38.797254 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1"} Mar 10 01:31:38.800411 kubelet[2683]: E0310 01:31:38.799979 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:38.800411 kubelet[2683]: E0310 01:31:38.800345 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-2d4ds" podUID="ea8ee6c7-d14f-40b8-b33b-78317f32f0c3" Mar 10 01:31:38.813584 containerd[1507]: time="2026-03-10T01:31:38.813169454Z" level=error msg="StopPodSandbox for \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\" failed" error="failed to destroy network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.815450 kubelet[2683]: E0310 01:31:38.815112 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:31:38.815450 kubelet[2683]: E0310 01:31:38.815202 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0"} Mar 10 01:31:38.816286 kubelet[2683]: E0310 01:31:38.815722 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"c035353a-3dfd-42e9-a291-75ad204d0496\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:38.816286 kubelet[2683]: E0310 01:31:38.815824 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"c035353a-3dfd-42e9-a291-75ad204d0496\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-794866df9b-424tn" podUID="c035353a-3dfd-42e9-a291-75ad204d0496" Mar 10 01:31:38.896479 containerd[1507]: time="2026-03-10T01:31:38.896422454Z" level=error msg="StopPodSandbox for \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\" failed" error="failed to destroy network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.898167 kubelet[2683]: E0310 01:31:38.897471 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:31:38.898167 kubelet[2683]: E0310 01:31:38.897585 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394"} Mar 10 01:31:38.898167 kubelet[2683]: E0310 01:31:38.897830 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"fa9659b6-52d9-4a4d-91d1-6f294264f5ee\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:38.898167 kubelet[2683]: E0310 01:31:38.897916 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"fa9659b6-52d9-4a4d-91d1-6f294264f5ee\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-apiserver-794866df9b-h52bq" podUID="fa9659b6-52d9-4a4d-91d1-6f294264f5ee" Mar 10 01:31:38.917471 containerd[1507]: time="2026-03-10T01:31:38.916505642Z" level=error msg="StopPodSandbox for \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\" failed" error="failed to destroy network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.917873 kubelet[2683]: E0310 01:31:38.917044 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:31:38.917873 kubelet[2683]: E0310 01:31:38.917547 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3"} Mar 10 01:31:38.917873 kubelet[2683]: E0310 01:31:38.917647 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8c562e47-a863-4e0d-8c19-218349a688f3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:38.918312 kubelet[2683]: E0310 01:31:38.917740 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8c562e47-a863-4e0d-8c19-218349a688f3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-jjvj8" podUID="8c562e47-a863-4e0d-8c19-218349a688f3" Mar 10 01:31:38.951894 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394-shm.mount: Deactivated successfully. Mar 10 01:31:38.952060 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3-shm.mount: Deactivated successfully. Mar 10 01:31:38.953043 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0-shm.mount: Deactivated successfully. Mar 10 01:31:38.998528 containerd[1507]: time="2026-03-10T01:31:38.998182025Z" level=error msg="StopPodSandbox for \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\" failed" error="failed to destroy network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.998528 containerd[1507]: time="2026-03-10T01:31:38.998379801Z" level=error msg="StopPodSandbox for \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\" failed" error="failed to destroy network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:38.999383 kubelet[2683]: E0310 01:31:38.998833 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:31:38.999383 kubelet[2683]: E0310 01:31:38.998976 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:38.999383 kubelet[2683]: E0310 01:31:38.999099 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8"} Mar 10 01:31:38.999383 kubelet[2683]: E0310 01:31:38.999179 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"7724f152-38e8-4ee2-b7cd-2755c7f51836\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:38.999383 kubelet[2683]: E0310 01:31:38.999030 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be"} Mar 10 01:31:38.999906 kubelet[2683]: E0310 01:31:38.999255 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"4c86706d-5198-4966-aa4f-480259ec456a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:38.999906 kubelet[2683]: E0310 01:31:38.999308 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"4c86706d-5198-4966-aa4f-480259ec456a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-cccfbd5cf-sxw9t" podUID="4c86706d-5198-4966-aa4f-480259ec456a" Mar 10 01:31:39.004118 kubelet[2683]: E0310 01:31:39.000301 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"7724f152-38e8-4ee2-b7cd-2755c7f51836\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7475b8444d-2s95h" podUID="7724f152-38e8-4ee2-b7cd-2755c7f51836" Mar 10 01:31:39.037078 containerd[1507]: time="2026-03-10T01:31:39.035868263Z" level=error msg="StopPodSandbox for \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\" failed" error="failed to destroy network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:39.040320 kubelet[2683]: E0310 01:31:39.036438 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:31:39.040320 kubelet[2683]: E0310 01:31:39.036515 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1"} Mar 10 01:31:39.040320 kubelet[2683]: E0310 01:31:39.036573 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2c9450c8-37a4-4965-96d3-67fee0cc59e8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:39.040320 kubelet[2683]: E0310 01:31:39.036648 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2c9450c8-37a4-4965-96d3-67fee0cc59e8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-77f6556c5c-68b2f" podUID="2c9450c8-37a4-4965-96d3-67fee0cc59e8" Mar 10 01:31:39.075887 containerd[1507]: time="2026-03-10T01:31:39.075822602Z" level=error msg="StopPodSandbox for \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\" failed" error="failed to destroy network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 10 01:31:39.076991 kubelet[2683]: E0310 01:31:39.076908 2683 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:31:39.077125 kubelet[2683]: E0310 01:31:39.077050 2683 kuberuntime_manager.go:1665] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f"} Mar 10 01:31:39.077125 kubelet[2683]: E0310 01:31:39.077115 2683 kuberuntime_manager.go:1233] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"71ea22a7-e0cb-4bd4-871f-854f4db6a991\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Mar 10 01:31:39.077341 kubelet[2683]: E0310 01:31:39.077167 2683 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"71ea22a7-e0cb-4bd4-871f-854f4db6a991\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-fhcl2" podUID="71ea22a7-e0cb-4bd4-871f-854f4db6a991" Mar 10 01:31:39.687052 containerd[1507]: time="2026-03-10T01:31:39.686400538Z" level=info msg="StopPodSandbox for \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\"" Mar 10 01:31:39.756937 systemd[1]: run-containerd-runc-k8s.io-828b65ed9f2fa67d61efbbe6be7fe5256bc16a55d277a6b804b27a98526d1147-runc.DiawrM.mount: Deactivated successfully. Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.827 [INFO][3951] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.828 [INFO][3951] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" iface="eth0" netns="/var/run/netns/cni-3496c744-1755-0bc3-50cf-653665565bed" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.829 [INFO][3951] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" iface="eth0" netns="/var/run/netns/cni-3496c744-1755-0bc3-50cf-653665565bed" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.830 [INFO][3951] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" iface="eth0" netns="/var/run/netns/cni-3496c744-1755-0bc3-50cf-653665565bed" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.830 [INFO][3951] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.830 [INFO][3951] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.896 [INFO][3978] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.898 [INFO][3978] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.898 [INFO][3978] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.908 [WARNING][3978] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.909 [INFO][3978] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.910 [INFO][3978] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:39.917996 containerd[1507]: 2026-03-10 01:31:39.915 [INFO][3951] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:39.919376 containerd[1507]: time="2026-03-10T01:31:39.919311106Z" level=info msg="TearDown network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\" successfully" Mar 10 01:31:39.919563 containerd[1507]: time="2026-03-10T01:31:39.919527640Z" level=info msg="StopPodSandbox for \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\" returns successfully" Mar 10 01:31:39.924481 systemd[1]: run-netns-cni\x2d3496c744\x2d1755\x2d0bc3\x2d50cf\x2d653665565bed.mount: Deactivated successfully. Mar 10 01:31:40.081344 kubelet[2683]: I0310 01:31:40.080120 2683 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-nginx-config\") pod \"7724f152-38e8-4ee2-b7cd-2755c7f51836\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " Mar 10 01:31:40.081344 kubelet[2683]: I0310 01:31:40.080186 2683 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4hft\" (UniqueName: \"kubernetes.io/projected/7724f152-38e8-4ee2-b7cd-2755c7f51836-kube-api-access-n4hft\") pod \"7724f152-38e8-4ee2-b7cd-2755c7f51836\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " Mar 10 01:31:40.081344 kubelet[2683]: I0310 01:31:40.080329 2683 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-backend-key-pair\") pod \"7724f152-38e8-4ee2-b7cd-2755c7f51836\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " Mar 10 01:31:40.081344 kubelet[2683]: I0310 01:31:40.080386 2683 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-ca-bundle\") pod \"7724f152-38e8-4ee2-b7cd-2755c7f51836\" (UID: \"7724f152-38e8-4ee2-b7cd-2755c7f51836\") " Mar 10 01:31:40.091107 kubelet[2683]: I0310 01:31:40.090833 2683 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-nginx-config" (OuterVolumeSpecName: "nginx-config") pod "7724f152-38e8-4ee2-b7cd-2755c7f51836" (UID: "7724f152-38e8-4ee2-b7cd-2755c7f51836"). InnerVolumeSpecName "nginx-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 10 01:31:40.091560 kubelet[2683]: I0310 01:31:40.088724 2683 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "7724f152-38e8-4ee2-b7cd-2755c7f51836" (UID: "7724f152-38e8-4ee2-b7cd-2755c7f51836"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 10 01:31:40.100311 kubelet[2683]: I0310 01:31:40.099943 2683 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7724f152-38e8-4ee2-b7cd-2755c7f51836-kube-api-access-n4hft" (OuterVolumeSpecName: "kube-api-access-n4hft") pod "7724f152-38e8-4ee2-b7cd-2755c7f51836" (UID: "7724f152-38e8-4ee2-b7cd-2755c7f51836"). InnerVolumeSpecName "kube-api-access-n4hft". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 10 01:31:40.101942 systemd[1]: var-lib-kubelet-pods-7724f152\x2d38e8\x2d4ee2\x2db7cd\x2d2755c7f51836-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dn4hft.mount: Deactivated successfully. Mar 10 01:31:40.102565 kubelet[2683]: I0310 01:31:40.102056 2683 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "7724f152-38e8-4ee2-b7cd-2755c7f51836" (UID: "7724f152-38e8-4ee2-b7cd-2755c7f51836"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 10 01:31:40.102106 systemd[1]: var-lib-kubelet-pods-7724f152\x2d38e8\x2d4ee2\x2db7cd\x2d2755c7f51836-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Mar 10 01:31:40.181040 kubelet[2683]: I0310 01:31:40.180957 2683 reconciler_common.go:299] "Volume detached for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-nginx-config\") on node \"srv-eb8wl.gb1.brightbox.com\" DevicePath \"\"" Mar 10 01:31:40.181040 kubelet[2683]: I0310 01:31:40.181025 2683 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-n4hft\" (UniqueName: \"kubernetes.io/projected/7724f152-38e8-4ee2-b7cd-2755c7f51836-kube-api-access-n4hft\") on node \"srv-eb8wl.gb1.brightbox.com\" DevicePath \"\"" Mar 10 01:31:40.181439 kubelet[2683]: I0310 01:31:40.181078 2683 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-backend-key-pair\") on node \"srv-eb8wl.gb1.brightbox.com\" DevicePath \"\"" Mar 10 01:31:40.181439 kubelet[2683]: I0310 01:31:40.181095 2683 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7724f152-38e8-4ee2-b7cd-2755c7f51836-whisker-ca-bundle\") on node \"srv-eb8wl.gb1.brightbox.com\" DevicePath \"\"" Mar 10 01:31:40.243147 systemd[1]: Removed slice kubepods-besteffort-pod7724f152_38e8_4ee2_b7cd_2755c7f51836.slice - libcontainer container kubepods-besteffort-pod7724f152_38e8_4ee2_b7cd_2755c7f51836.slice. Mar 10 01:31:40.830544 systemd[1]: run-containerd-runc-k8s.io-828b65ed9f2fa67d61efbbe6be7fe5256bc16a55d277a6b804b27a98526d1147-runc.jktFmQ.mount: Deactivated successfully. Mar 10 01:31:41.002383 systemd[1]: Created slice kubepods-besteffort-podd978a4b9_ea49_43cc_a0a9_0823c44d34e2.slice - libcontainer container kubepods-besteffort-podd978a4b9_ea49_43cc_a0a9_0823c44d34e2.slice. Mar 10 01:31:41.090374 kubelet[2683]: I0310 01:31:41.089306 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kfsq\" (UniqueName: \"kubernetes.io/projected/d978a4b9-ea49-43cc-a0a9-0823c44d34e2-kube-api-access-9kfsq\") pod \"whisker-5d545d8c69-b54r6\" (UID: \"d978a4b9-ea49-43cc-a0a9-0823c44d34e2\") " pod="calico-system/whisker-5d545d8c69-b54r6" Mar 10 01:31:41.090374 kubelet[2683]: I0310 01:31:41.089422 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-config\" (UniqueName: \"kubernetes.io/configmap/d978a4b9-ea49-43cc-a0a9-0823c44d34e2-nginx-config\") pod \"whisker-5d545d8c69-b54r6\" (UID: \"d978a4b9-ea49-43cc-a0a9-0823c44d34e2\") " pod="calico-system/whisker-5d545d8c69-b54r6" Mar 10 01:31:41.090374 kubelet[2683]: I0310 01:31:41.089474 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d978a4b9-ea49-43cc-a0a9-0823c44d34e2-whisker-backend-key-pair\") pod \"whisker-5d545d8c69-b54r6\" (UID: \"d978a4b9-ea49-43cc-a0a9-0823c44d34e2\") " pod="calico-system/whisker-5d545d8c69-b54r6" Mar 10 01:31:41.090374 kubelet[2683]: I0310 01:31:41.089507 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d978a4b9-ea49-43cc-a0a9-0823c44d34e2-whisker-ca-bundle\") pod \"whisker-5d545d8c69-b54r6\" (UID: \"d978a4b9-ea49-43cc-a0a9-0823c44d34e2\") " pod="calico-system/whisker-5d545d8c69-b54r6" Mar 10 01:31:41.314305 containerd[1507]: time="2026-03-10T01:31:41.312523611Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d545d8c69-b54r6,Uid:d978a4b9-ea49-43cc-a0a9-0823c44d34e2,Namespace:calico-system,Attempt:0,}" Mar 10 01:31:41.617289 kernel: calico-node[4102]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Mar 10 01:31:41.648368 systemd-networkd[1419]: caliadc0a6c380a: Link UP Mar 10 01:31:41.649413 systemd-networkd[1419]: caliadc0a6c380a: Gained carrier Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.444 [ERROR][4119] cni-plugin/utils.go 116: File does not exist, skipping the error since RequireMTUFile is false error=open /var/lib/calico/mtu: no such file or directory filename="/var/lib/calico/mtu" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.465 [INFO][4119] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0 whisker-5d545d8c69- calico-system d978a4b9-ea49-43cc-a0a9-0823c44d34e2 957 0 2026-03-10 01:31:40 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:5d545d8c69 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com whisker-5d545d8c69-b54r6 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] caliadc0a6c380a [] [] }} ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.465 [INFO][4119] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.532 [INFO][4141] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" HandleID="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.551 [INFO][4141] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" HandleID="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002fb9b0), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"whisker-5d545d8c69-b54r6", "timestamp":"2026-03-10 01:31:41.532637015 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000248580)} Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.551 [INFO][4141] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.552 [INFO][4141] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.552 [INFO][4141] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.558 [INFO][4141] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.566 [INFO][4141] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.575 [INFO][4141] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.578 [INFO][4141] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.584 [INFO][4141] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.584 [INFO][4141] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.587 [INFO][4141] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701 Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.597 [INFO][4141] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.604 [INFO][4141] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.65/26] block=192.168.57.64/26 handle="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.605 [INFO][4141] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.65/26] handle="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.605 [INFO][4141] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:41.738972 containerd[1507]: 2026-03-10 01:31:41.605 [INFO][4141] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.65/26] IPv6=[] ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" HandleID="k8s-pod-network.58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" Mar 10 01:31:41.741198 containerd[1507]: 2026-03-10 01:31:41.610 [INFO][4119] cni-plugin/k8s.go 418: Populated endpoint ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0", GenerateName:"whisker-5d545d8c69-", Namespace:"calico-system", SelfLink:"", UID:"d978a4b9-ea49-43cc-a0a9-0823c44d34e2", ResourceVersion:"957", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5d545d8c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"whisker-5d545d8c69-b54r6", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.57.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"caliadc0a6c380a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:41.741198 containerd[1507]: 2026-03-10 01:31:41.610 [INFO][4119] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.65/32] ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" Mar 10 01:31:41.741198 containerd[1507]: 2026-03-10 01:31:41.610 [INFO][4119] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliadc0a6c380a ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" Mar 10 01:31:41.741198 containerd[1507]: 2026-03-10 01:31:41.650 [INFO][4119] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" Mar 10 01:31:41.741198 containerd[1507]: 2026-03-10 01:31:41.660 [INFO][4119] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0", GenerateName:"whisker-5d545d8c69-", Namespace:"calico-system", SelfLink:"", UID:"d978a4b9-ea49-43cc-a0a9-0823c44d34e2", ResourceVersion:"957", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5d545d8c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701", Pod:"whisker-5d545d8c69-b54r6", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.57.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"caliadc0a6c380a", MAC:"f2:db:92:bc:1f:45", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:41.741198 containerd[1507]: 2026-03-10 01:31:41.730 [INFO][4119] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701" Namespace="calico-system" Pod="whisker-5d545d8c69-b54r6" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--5d545d8c69--b54r6-eth0" Mar 10 01:31:42.215466 containerd[1507]: time="2026-03-10T01:31:42.206883900Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:42.215838 containerd[1507]: time="2026-03-10T01:31:42.215317673Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:42.215838 containerd[1507]: time="2026-03-10T01:31:42.215493665Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:42.216470 containerd[1507]: time="2026-03-10T01:31:42.215994672Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:42.289363 kubelet[2683]: I0310 01:31:42.244729 2683 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7724f152-38e8-4ee2-b7cd-2755c7f51836" path="/var/lib/kubelet/pods/7724f152-38e8-4ee2-b7cd-2755c7f51836/volumes" Mar 10 01:31:42.370485 systemd[1]: Started cri-containerd-58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701.scope - libcontainer container 58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701. Mar 10 01:31:42.619678 containerd[1507]: time="2026-03-10T01:31:42.619606254Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d545d8c69-b54r6,Uid:d978a4b9-ea49-43cc-a0a9-0823c44d34e2,Namespace:calico-system,Attempt:0,} returns sandbox id \"58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701\"" Mar 10 01:31:42.759608 containerd[1507]: time="2026-03-10T01:31:42.759548059Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\"" Mar 10 01:31:43.088562 systemd-networkd[1419]: caliadc0a6c380a: Gained IPv6LL Mar 10 01:31:43.175823 systemd-networkd[1419]: vxlan.calico: Link UP Mar 10 01:31:43.175837 systemd-networkd[1419]: vxlan.calico: Gained carrier Mar 10 01:31:44.304846 systemd-networkd[1419]: vxlan.calico: Gained IPv6LL Mar 10 01:31:44.723839 containerd[1507]: time="2026-03-10T01:31:44.722493786Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.31.4: active requests=0, bytes read=6039889" Mar 10 01:31:44.726169 containerd[1507]: time="2026-03-10T01:31:44.726058404Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.31.4\" with image id \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\", size \"7595926\" in 1.966444165s" Mar 10 01:31:44.726169 containerd[1507]: time="2026-03-10T01:31:44.726113710Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.31.4\" returns image reference \"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\"" Mar 10 01:31:44.731207 containerd[1507]: time="2026-03-10T01:31:44.731075788Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:44.740001 containerd[1507]: time="2026-03-10T01:31:44.738900036Z" level=info msg="ImageCreate event name:\"sha256:c02b0051502f3aa7f0815d838ea93b53dfb6bd13f185d229260e08200daf7cf7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:44.741796 containerd[1507]: time="2026-03-10T01:31:44.741701754Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:9690cd395efad501f2e0c40ce4969d87b736ae2e5ed454644e7b0fd8f756bfbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:44.766166 containerd[1507]: time="2026-03-10T01:31:44.766114553Z" level=info msg="CreateContainer within sandbox \"58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Mar 10 01:31:44.818531 containerd[1507]: time="2026-03-10T01:31:44.818447600Z" level=info msg="CreateContainer within sandbox \"58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"35f941a5c8f6f54a8b782f1d6ec74c0e7971f3d780d52500be3731675a05a788\"" Mar 10 01:31:44.821461 containerd[1507]: time="2026-03-10T01:31:44.821340879Z" level=info msg="StartContainer for \"35f941a5c8f6f54a8b782f1d6ec74c0e7971f3d780d52500be3731675a05a788\"" Mar 10 01:31:44.896549 systemd[1]: Started cri-containerd-35f941a5c8f6f54a8b782f1d6ec74c0e7971f3d780d52500be3731675a05a788.scope - libcontainer container 35f941a5c8f6f54a8b782f1d6ec74c0e7971f3d780d52500be3731675a05a788. Mar 10 01:31:44.963760 containerd[1507]: time="2026-03-10T01:31:44.963631058Z" level=info msg="StartContainer for \"35f941a5c8f6f54a8b782f1d6ec74c0e7971f3d780d52500be3731675a05a788\" returns successfully" Mar 10 01:31:44.974418 containerd[1507]: time="2026-03-10T01:31:44.974167879Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\"" Mar 10 01:31:47.180497 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount301505628.mount: Deactivated successfully. Mar 10 01:31:47.197386 containerd[1507]: time="2026-03-10T01:31:47.197335317Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:47.198481 containerd[1507]: time="2026-03-10T01:31:47.198085577Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.31.4: active requests=0, bytes read=17609475" Mar 10 01:31:47.206222 containerd[1507]: time="2026-03-10T01:31:47.206152743Z" level=info msg="ImageCreate event name:\"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:47.217241 containerd[1507]: time="2026-03-10T01:31:47.215616648Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:47.217241 containerd[1507]: time="2026-03-10T01:31:47.216887842Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" with image id \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:d252061aa298c4b17cf092517b5126af97cf95e0f56b21281b95a5f8702f15fc\", size \"17609305\" in 2.242675014s" Mar 10 01:31:47.217241 containerd[1507]: time="2026-03-10T01:31:47.216926280Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.31.4\" returns image reference \"sha256:0749e3da0398e8402eb119f09acf145e5dd9759adb6eb3802ad6dc1b9bbedf1c\"" Mar 10 01:31:47.231667 containerd[1507]: time="2026-03-10T01:31:47.231600979Z" level=info msg="CreateContainer within sandbox \"58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Mar 10 01:31:47.248828 containerd[1507]: time="2026-03-10T01:31:47.248633013Z" level=info msg="CreateContainer within sandbox \"58058ce747176a0eef9355bc393a9810503e4454a6c6db33a79ede6ad818d701\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"a93bdd69357fdba0f23361d3d720d83d483a2d96da7fd5938ff898aa9984c822\"" Mar 10 01:31:47.249558 containerd[1507]: time="2026-03-10T01:31:47.249357403Z" level=info msg="StartContainer for \"a93bdd69357fdba0f23361d3d720d83d483a2d96da7fd5938ff898aa9984c822\"" Mar 10 01:31:47.309441 systemd[1]: Started cri-containerd-a93bdd69357fdba0f23361d3d720d83d483a2d96da7fd5938ff898aa9984c822.scope - libcontainer container a93bdd69357fdba0f23361d3d720d83d483a2d96da7fd5938ff898aa9984c822. Mar 10 01:31:47.381814 containerd[1507]: time="2026-03-10T01:31:47.381729333Z" level=info msg="StartContainer for \"a93bdd69357fdba0f23361d3d720d83d483a2d96da7fd5938ff898aa9984c822\" returns successfully" Mar 10 01:31:47.889609 kubelet[2683]: I0310 01:31:47.887713 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-5d545d8c69-b54r6" podStartSLOduration=3.334731779 podStartE2EDuration="7.887032965s" podCreationTimestamp="2026-03-10 01:31:40 +0000 UTC" firstStartedPulling="2026-03-10 01:31:42.666585319 +0000 UTC m=+52.774068427" lastFinishedPulling="2026-03-10 01:31:47.218886505 +0000 UTC m=+57.326369613" observedRunningTime="2026-03-10 01:31:47.884673433 +0000 UTC m=+57.992156555" watchObservedRunningTime="2026-03-10 01:31:47.887032965 +0000 UTC m=+57.994516071" Mar 10 01:31:50.180112 containerd[1507]: time="2026-03-10T01:31:50.179502190Z" level=info msg="StopPodSandbox for \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\"" Mar 10 01:31:50.232848 containerd[1507]: time="2026-03-10T01:31:50.231582998Z" level=info msg="StopPodSandbox for \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\"" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.524 [INFO][4427] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.524 [INFO][4427] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" iface="eth0" netns="/var/run/netns/cni-9b150cde-ddbd-25be-3e64-06d6ad0173d8" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.525 [INFO][4427] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" iface="eth0" netns="/var/run/netns/cni-9b150cde-ddbd-25be-3e64-06d6ad0173d8" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.526 [INFO][4427] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" iface="eth0" netns="/var/run/netns/cni-9b150cde-ddbd-25be-3e64-06d6ad0173d8" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.527 [INFO][4427] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.527 [INFO][4427] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.699 [INFO][4443] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.699 [INFO][4443] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.700 [INFO][4443] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.715 [WARNING][4443] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.715 [INFO][4443] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.718 [INFO][4443] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:50.730122 containerd[1507]: 2026-03-10 01:31:50.725 [INFO][4427] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:31:50.736023 containerd[1507]: time="2026-03-10T01:31:50.731418258Z" level=info msg="TearDown network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\" successfully" Mar 10 01:31:50.736023 containerd[1507]: time="2026-03-10T01:31:50.731678390Z" level=info msg="StopPodSandbox for \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\" returns successfully" Mar 10 01:31:50.738973 systemd[1]: run-netns-cni\x2d9b150cde\x2dddbd\x2d25be\x2d3e64\x2d06d6ad0173d8.mount: Deactivated successfully. Mar 10 01:31:50.745420 containerd[1507]: time="2026-03-10T01:31:50.745248012Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-cccfbd5cf-sxw9t,Uid:4c86706d-5198-4966-aa4f-480259ec456a,Namespace:calico-system,Attempt:1,}" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.482 [WARNING][4414] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.484 [INFO][4414] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.484 [INFO][4414] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" iface="eth0" netns="" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.484 [INFO][4414] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.484 [INFO][4414] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.705 [INFO][4438] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.707 [INFO][4438] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.718 [INFO][4438] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.735 [WARNING][4438] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.735 [INFO][4438] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.743 [INFO][4438] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:50.752506 containerd[1507]: 2026-03-10 01:31:50.747 [INFO][4414] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.753898 containerd[1507]: time="2026-03-10T01:31:50.753285637Z" level=info msg="TearDown network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\" successfully" Mar 10 01:31:50.753898 containerd[1507]: time="2026-03-10T01:31:50.753315910Z" level=info msg="StopPodSandbox for \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\" returns successfully" Mar 10 01:31:50.758692 containerd[1507]: time="2026-03-10T01:31:50.758654464Z" level=info msg="RemovePodSandbox for \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\"" Mar 10 01:31:50.759064 containerd[1507]: time="2026-03-10T01:31:50.758848676Z" level=info msg="Forcibly stopping sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\"" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.837 [WARNING][4467] cni-plugin/k8s.go 610: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.838 [INFO][4467] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.838 [INFO][4467] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" iface="eth0" netns="" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.838 [INFO][4467] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.838 [INFO][4467] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.935 [INFO][4477] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.936 [INFO][4477] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.936 [INFO][4477] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.948 [WARNING][4477] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.949 [INFO][4477] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" HandleID="k8s-pod-network.c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Workload="srv--eb8wl.gb1.brightbox.com-k8s-whisker--7475b8444d--2s95h-eth0" Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.952 [INFO][4477] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:50.957507 containerd[1507]: 2026-03-10 01:31:50.954 [INFO][4467] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8" Mar 10 01:31:50.958803 containerd[1507]: time="2026-03-10T01:31:50.957914502Z" level=info msg="TearDown network for sandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\" successfully" Mar 10 01:31:50.968395 containerd[1507]: time="2026-03-10T01:31:50.968171653Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:31:50.968395 containerd[1507]: time="2026-03-10T01:31:50.968317165Z" level=info msg="RemovePodSandbox \"c2b105c5de6b510dc7e69ca22d328fa1f30ae1f4840107447c621a7005a74dd8\" returns successfully" Mar 10 01:31:51.008984 systemd-networkd[1419]: calid991037db6e: Link UP Mar 10 01:31:51.011552 systemd-networkd[1419]: calid991037db6e: Gained carrier Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.855 [INFO][4456] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0 goldmane-cccfbd5cf- calico-system 4c86706d-5198-4966-aa4f-480259ec456a 1001 0 2026-03-10 01:31:11 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:cccfbd5cf projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com goldmane-cccfbd5cf-sxw9t eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calid991037db6e [] [] }} ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.856 [INFO][4456] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.935 [INFO][4483] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" HandleID="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.953 [INFO][4483] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" HandleID="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000103cc0), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"goldmane-cccfbd5cf-sxw9t", "timestamp":"2026-03-10 01:31:50.935807413 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00027b080)} Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.953 [INFO][4483] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.954 [INFO][4483] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.954 [INFO][4483] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.960 [INFO][4483] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.968 [INFO][4483] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.974 [INFO][4483] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.978 [INFO][4483] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.981 [INFO][4483] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.981 [INFO][4483] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.982 [INFO][4483] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42 Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.989 [INFO][4483] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.996 [INFO][4483] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.66/26] block=192.168.57.64/26 handle="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.996 [INFO][4483] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.66/26] handle="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.997 [INFO][4483] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:51.046439 containerd[1507]: 2026-03-10 01:31:50.997 [INFO][4483] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.66/26] IPv6=[] ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" HandleID="k8s-pod-network.c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:51.050111 containerd[1507]: 2026-03-10 01:31:51.000 [INFO][4456] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0", GenerateName:"goldmane-cccfbd5cf-", Namespace:"calico-system", SelfLink:"", UID:"4c86706d-5198-4966-aa4f-480259ec456a", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"cccfbd5cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"goldmane-cccfbd5cf-sxw9t", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.57.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calid991037db6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:51.050111 containerd[1507]: 2026-03-10 01:31:51.001 [INFO][4456] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.66/32] ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:51.050111 containerd[1507]: 2026-03-10 01:31:51.001 [INFO][4456] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid991037db6e ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:51.050111 containerd[1507]: 2026-03-10 01:31:51.007 [INFO][4456] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:51.050111 containerd[1507]: 2026-03-10 01:31:51.008 [INFO][4456] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0", GenerateName:"goldmane-cccfbd5cf-", Namespace:"calico-system", SelfLink:"", UID:"4c86706d-5198-4966-aa4f-480259ec456a", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"cccfbd5cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42", Pod:"goldmane-cccfbd5cf-sxw9t", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.57.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calid991037db6e", MAC:"ae:60:29:d1:7d:10", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:51.050111 containerd[1507]: 2026-03-10 01:31:51.038 [INFO][4456] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42" Namespace="calico-system" Pod="goldmane-cccfbd5cf-sxw9t" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:31:51.086337 containerd[1507]: time="2026-03-10T01:31:51.084121379Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:51.086337 containerd[1507]: time="2026-03-10T01:31:51.084291622Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:51.086337 containerd[1507]: time="2026-03-10T01:31:51.084325599Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:51.086337 containerd[1507]: time="2026-03-10T01:31:51.084472580Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:51.148292 systemd[1]: Started cri-containerd-c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42.scope - libcontainer container c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42. Mar 10 01:31:51.232814 containerd[1507]: time="2026-03-10T01:31:51.232577144Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-cccfbd5cf-sxw9t,Uid:4c86706d-5198-4966-aa4f-480259ec456a,Namespace:calico-system,Attempt:1,} returns sandbox id \"c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42\"" Mar 10 01:31:51.240398 containerd[1507]: time="2026-03-10T01:31:51.239598138Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\"" Mar 10 01:31:52.234513 containerd[1507]: time="2026-03-10T01:31:52.233996358Z" level=info msg="StopPodSandbox for \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\"" Mar 10 01:31:52.238257 containerd[1507]: time="2026-03-10T01:31:52.236373768Z" level=info msg="StopPodSandbox for \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\"" Mar 10 01:31:52.242239 containerd[1507]: time="2026-03-10T01:31:52.242167984Z" level=info msg="StopPodSandbox for \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\"" Mar 10 01:31:52.245541 containerd[1507]: time="2026-03-10T01:31:52.244566458Z" level=info msg="StopPodSandbox for \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\"" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.426 [INFO][4596] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.427 [INFO][4596] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" iface="eth0" netns="/var/run/netns/cni-017c61f9-cc92-37de-7cb8-6affe7ac3887" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.428 [INFO][4596] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" iface="eth0" netns="/var/run/netns/cni-017c61f9-cc92-37de-7cb8-6affe7ac3887" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.430 [INFO][4596] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" iface="eth0" netns="/var/run/netns/cni-017c61f9-cc92-37de-7cb8-6affe7ac3887" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.430 [INFO][4596] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.430 [INFO][4596] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.597 [INFO][4613] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.597 [INFO][4613] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.598 [INFO][4613] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.630 [WARNING][4613] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.631 [INFO][4613] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.636 [INFO][4613] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:52.653427 containerd[1507]: 2026-03-10 01:31:52.647 [INFO][4596] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:31:52.657323 containerd[1507]: time="2026-03-10T01:31:52.657168777Z" level=info msg="TearDown network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\" successfully" Mar 10 01:31:52.659312 containerd[1507]: time="2026-03-10T01:31:52.659278966Z" level=info msg="StopPodSandbox for \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\" returns successfully" Mar 10 01:31:52.666879 containerd[1507]: time="2026-03-10T01:31:52.666743932Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-jjvj8,Uid:8c562e47-a863-4e0d-8c19-218349a688f3,Namespace:kube-system,Attempt:1,}" Mar 10 01:31:52.667201 systemd[1]: run-netns-cni\x2d017c61f9\x2dcc92\x2d37de\x2d7cb8\x2d6affe7ac3887.mount: Deactivated successfully. Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.521 [INFO][4584] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.525 [INFO][4584] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" iface="eth0" netns="/var/run/netns/cni-10260f22-ee0c-2dc8-a383-692d636250e5" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.525 [INFO][4584] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" iface="eth0" netns="/var/run/netns/cni-10260f22-ee0c-2dc8-a383-692d636250e5" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.526 [INFO][4584] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" iface="eth0" netns="/var/run/netns/cni-10260f22-ee0c-2dc8-a383-692d636250e5" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.526 [INFO][4584] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.526 [INFO][4584] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.624 [INFO][4628] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.624 [INFO][4628] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.638 [INFO][4628] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.669 [WARNING][4628] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.670 [INFO][4628] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.674 [INFO][4628] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:52.696279 containerd[1507]: 2026-03-10 01:31:52.680 [INFO][4584] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:31:52.701012 containerd[1507]: time="2026-03-10T01:31:52.693833756Z" level=info msg="TearDown network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\" successfully" Mar 10 01:31:52.702970 containerd[1507]: time="2026-03-10T01:31:52.701191234Z" level=info msg="StopPodSandbox for \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\" returns successfully" Mar 10 01:31:52.702811 systemd[1]: run-netns-cni\x2d10260f22\x2dee0c\x2d2dc8\x2da383\x2d692d636250e5.mount: Deactivated successfully. Mar 10 01:31:52.712977 containerd[1507]: time="2026-03-10T01:31:52.711454388Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fhcl2,Uid:71ea22a7-e0cb-4bd4-871f-854f4db6a991,Namespace:calico-system,Attempt:1,}" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.498 [INFO][4597] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.499 [INFO][4597] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" iface="eth0" netns="/var/run/netns/cni-33d29542-bdb0-293e-f7ca-45b7807000e0" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.501 [INFO][4597] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" iface="eth0" netns="/var/run/netns/cni-33d29542-bdb0-293e-f7ca-45b7807000e0" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.502 [INFO][4597] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" iface="eth0" netns="/var/run/netns/cni-33d29542-bdb0-293e-f7ca-45b7807000e0" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.502 [INFO][4597] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.502 [INFO][4597] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.638 [INFO][4621] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.647 [INFO][4621] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.674 [INFO][4621] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.698 [WARNING][4621] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.707 [INFO][4621] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.717 [INFO][4621] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:52.759582 containerd[1507]: 2026-03-10 01:31:52.725 [INFO][4597] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:31:52.769043 systemd[1]: run-netns-cni\x2d33d29542\x2dbdb0\x2d293e\x2df7ca\x2d45b7807000e0.mount: Deactivated successfully. Mar 10 01:31:52.772347 containerd[1507]: time="2026-03-10T01:31:52.771458668Z" level=info msg="TearDown network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\" successfully" Mar 10 01:31:52.772347 containerd[1507]: time="2026-03-10T01:31:52.771529138Z" level=info msg="StopPodSandbox for \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\" returns successfully" Mar 10 01:31:52.775547 containerd[1507]: time="2026-03-10T01:31:52.775499602Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-h52bq,Uid:fa9659b6-52d9-4a4d-91d1-6f294264f5ee,Namespace:calico-system,Attempt:1,}" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.578 [INFO][4598] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.579 [INFO][4598] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" iface="eth0" netns="/var/run/netns/cni-b59a4d3c-40a1-6ead-5ce7-dbe52cef379f" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.580 [INFO][4598] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" iface="eth0" netns="/var/run/netns/cni-b59a4d3c-40a1-6ead-5ce7-dbe52cef379f" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.580 [INFO][4598] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" iface="eth0" netns="/var/run/netns/cni-b59a4d3c-40a1-6ead-5ce7-dbe52cef379f" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.580 [INFO][4598] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.581 [INFO][4598] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.790 [INFO][4634] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.791 [INFO][4634] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.791 [INFO][4634] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.814 [WARNING][4634] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.815 [INFO][4634] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.818 [INFO][4634] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:52.837014 containerd[1507]: 2026-03-10 01:31:52.831 [INFO][4598] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:31:52.840209 containerd[1507]: time="2026-03-10T01:31:52.839318202Z" level=info msg="TearDown network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\" successfully" Mar 10 01:31:52.840209 containerd[1507]: time="2026-03-10T01:31:52.839373626Z" level=info msg="StopPodSandbox for \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\" returns successfully" Mar 10 01:31:52.843729 containerd[1507]: time="2026-03-10T01:31:52.842601452Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-77f6556c5c-68b2f,Uid:2c9450c8-37a4-4965-96d3-67fee0cc59e8,Namespace:calico-system,Attempt:1,}" Mar 10 01:31:52.882290 systemd-networkd[1419]: calid991037db6e: Gained IPv6LL Mar 10 01:31:53.206384 systemd-networkd[1419]: cali64cc8014aee: Link UP Mar 10 01:31:53.211364 systemd-networkd[1419]: cali64cc8014aee: Gained carrier Mar 10 01:31:53.250462 containerd[1507]: time="2026-03-10T01:31:53.250402024Z" level=info msg="StopPodSandbox for \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\"" Mar 10 01:31:53.279252 containerd[1507]: time="2026-03-10T01:31:53.278822617Z" level=info msg="StopPodSandbox for \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\"" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:52.951 [INFO][4655] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0 calico-apiserver-794866df9b- calico-system fa9659b6-52d9-4a4d-91d1-6f294264f5ee 1016 0 2026-03-10 01:31:11 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:794866df9b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com calico-apiserver-794866df9b-h52bq eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] cali64cc8014aee [] [] }} ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:52.952 [INFO][4655] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.066 [INFO][4688] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" HandleID="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.088 [INFO][4688] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" HandleID="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fec0), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"calico-apiserver-794866df9b-h52bq", "timestamp":"2026-03-10 01:31:53.066824203 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc000520000)} Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.089 [INFO][4688] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.090 [INFO][4688] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.090 [INFO][4688] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.096 [INFO][4688] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.111 [INFO][4688] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.125 [INFO][4688] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.134 [INFO][4688] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.142 [INFO][4688] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.143 [INFO][4688] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.149 [INFO][4688] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14 Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.161 [INFO][4688] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.176 [INFO][4688] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.67/26] block=192.168.57.64/26 handle="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.176 [INFO][4688] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.67/26] handle="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.176 [INFO][4688] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:53.368985 containerd[1507]: 2026-03-10 01:31:53.176 [INFO][4688] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.67/26] IPv6=[] ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" HandleID="k8s-pod-network.50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:53.373666 containerd[1507]: 2026-03-10 01:31:53.185 [INFO][4655] cni-plugin/k8s.go 418: Populated endpoint ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"fa9659b6-52d9-4a4d-91d1-6f294264f5ee", ResourceVersion:"1016", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"calico-apiserver-794866df9b-h52bq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali64cc8014aee", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:53.373666 containerd[1507]: 2026-03-10 01:31:53.187 [INFO][4655] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.67/32] ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:53.373666 containerd[1507]: 2026-03-10 01:31:53.187 [INFO][4655] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali64cc8014aee ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:53.373666 containerd[1507]: 2026-03-10 01:31:53.239 [INFO][4655] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:53.373666 containerd[1507]: 2026-03-10 01:31:53.256 [INFO][4655] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"fa9659b6-52d9-4a4d-91d1-6f294264f5ee", ResourceVersion:"1016", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14", Pod:"calico-apiserver-794866df9b-h52bq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali64cc8014aee", MAC:"0e:79:98:be:ee:22", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:53.373666 containerd[1507]: 2026-03-10 01:31:53.359 [INFO][4655] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14" Namespace="calico-system" Pod="calico-apiserver-794866df9b-h52bq" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:31:53.489399 systemd-networkd[1419]: calia08e75d8c26: Link UP Mar 10 01:31:53.493789 systemd-networkd[1419]: calia08e75d8c26: Gained carrier Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.054 [INFO][4653] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0 csi-node-driver- calico-system 71ea22a7-e0cb-4bd4-871f-854f4db6a991 1017 0 2026-03-10 01:31:12 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:98cbb5577 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com csi-node-driver-fhcl2 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calia08e75d8c26 [] [] }} ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.055 [INFO][4653] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.244 [INFO][4702] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" HandleID="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.281 [INFO][4702] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" HandleID="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000608120), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"csi-node-driver-fhcl2", "timestamp":"2026-03-10 01:31:53.244311004 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00015e160)} Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.281 [INFO][4702] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.281 [INFO][4702] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.281 [INFO][4702] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.324 [INFO][4702] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.390 [INFO][4702] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.414 [INFO][4702] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.421 [INFO][4702] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.429 [INFO][4702] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.429 [INFO][4702] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.433 [INFO][4702] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.449 [INFO][4702] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.465 [INFO][4702] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.68/26] block=192.168.57.64/26 handle="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.469 [INFO][4702] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.68/26] handle="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.470 [INFO][4702] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:53.595282 containerd[1507]: 2026-03-10 01:31:53.470 [INFO][4702] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.68/26] IPv6=[] ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" HandleID="k8s-pod-network.b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:53.598465 containerd[1507]: 2026-03-10 01:31:53.478 [INFO][4653] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"71ea22a7-e0cb-4bd4-871f-854f4db6a991", ResourceVersion:"1017", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"98cbb5577", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"csi-node-driver-fhcl2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.57.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia08e75d8c26", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:53.598465 containerd[1507]: 2026-03-10 01:31:53.478 [INFO][4653] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.68/32] ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:53.598465 containerd[1507]: 2026-03-10 01:31:53.478 [INFO][4653] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia08e75d8c26 ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:53.598465 containerd[1507]: 2026-03-10 01:31:53.502 [INFO][4653] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:53.598465 containerd[1507]: 2026-03-10 01:31:53.526 [INFO][4653] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"71ea22a7-e0cb-4bd4-871f-854f4db6a991", ResourceVersion:"1017", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"98cbb5577", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb", Pod:"csi-node-driver-fhcl2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.57.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia08e75d8c26", MAC:"72:62:06:f3:b2:df", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:53.598465 containerd[1507]: 2026-03-10 01:31:53.570 [INFO][4653] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb" Namespace="calico-system" Pod="csi-node-driver-fhcl2" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:31:53.607241 containerd[1507]: time="2026-03-10T01:31:53.600403673Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:53.607241 containerd[1507]: time="2026-03-10T01:31:53.601147555Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:53.607241 containerd[1507]: time="2026-03-10T01:31:53.601187229Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:53.607241 containerd[1507]: time="2026-03-10T01:31:53.602728698Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:53.692271 systemd[1]: run-netns-cni\x2db59a4d3c\x2d40a1\x2d6ead\x2d5ce7\x2ddbe52cef379f.mount: Deactivated successfully. Mar 10 01:31:53.784665 systemd-networkd[1419]: calid615bcab3d4: Link UP Mar 10 01:31:53.787350 systemd-networkd[1419]: calid615bcab3d4: Gained carrier Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.000 [INFO][4642] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0 coredns-66bc5c9577- kube-system 8c562e47-a863-4e0d-8c19-218349a688f3 1015 0 2026-03-10 01:30:56 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com coredns-66bc5c9577-jjvj8 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calid615bcab3d4 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.000 [INFO][4642] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.190 [INFO][4695] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" HandleID="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.283 [INFO][4695] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" HandleID="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001039b0), Attrs:map[string]string{"namespace":"kube-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"coredns-66bc5c9577-jjvj8", "timestamp":"2026-03-10 01:31:53.190897563 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0002db1e0)} Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.284 [INFO][4695] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.471 [INFO][4695] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.471 [INFO][4695] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.481 [INFO][4695] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.506 [INFO][4695] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.573 [INFO][4695] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.599 [INFO][4695] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.618 [INFO][4695] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.619 [INFO][4695] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.633 [INFO][4695] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020 Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.648 [INFO][4695] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.696 [INFO][4695] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.69/26] block=192.168.57.64/26 handle="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.696 [INFO][4695] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.69/26] handle="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.696 [INFO][4695] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:53.870209 containerd[1507]: 2026-03-10 01:31:53.696 [INFO][4695] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.69/26] IPv6=[] ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" HandleID="k8s-pod-network.776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:53.871414 containerd[1507]: 2026-03-10 01:31:53.709 [INFO][4642] cni-plugin/k8s.go 418: Populated endpoint ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"8c562e47-a863-4e0d-8c19-218349a688f3", ResourceVersion:"1015", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"coredns-66bc5c9577-jjvj8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid615bcab3d4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:53.871414 containerd[1507]: 2026-03-10 01:31:53.721 [INFO][4642] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.69/32] ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:53.871414 containerd[1507]: 2026-03-10 01:31:53.721 [INFO][4642] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid615bcab3d4 ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:53.871414 containerd[1507]: 2026-03-10 01:31:53.805 [INFO][4642] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:53.871414 containerd[1507]: 2026-03-10 01:31:53.810 [INFO][4642] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"8c562e47-a863-4e0d-8c19-218349a688f3", ResourceVersion:"1015", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020", Pod:"coredns-66bc5c9577-jjvj8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid615bcab3d4", MAC:"a6:35:91:f6:da:6e", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:53.871830 containerd[1507]: 2026-03-10 01:31:53.844 [INFO][4642] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020" Namespace="kube-system" Pod="coredns-66bc5c9577-jjvj8" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:31:53.881504 systemd[1]: Started cri-containerd-50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14.scope - libcontainer container 50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14. Mar 10 01:31:54.008213 containerd[1507]: time="2026-03-10T01:31:53.991890260Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:54.008213 containerd[1507]: time="2026-03-10T01:31:53.991971288Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:54.008213 containerd[1507]: time="2026-03-10T01:31:53.991990248Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:54.008213 containerd[1507]: time="2026-03-10T01:31:53.992137246Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:54.033191 systemd-networkd[1419]: calib01fa7c8b48: Link UP Mar 10 01:31:54.038497 systemd-networkd[1419]: calib01fa7c8b48: Gained carrier Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.627 [INFO][4742] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.627 [INFO][4742] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" iface="eth0" netns="/var/run/netns/cni-f3ea6fa0-8bd4-7b3c-bcbf-8cc4cdf03de4" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.627 [INFO][4742] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" iface="eth0" netns="/var/run/netns/cni-f3ea6fa0-8bd4-7b3c-bcbf-8cc4cdf03de4" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.629 [INFO][4742] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" iface="eth0" netns="/var/run/netns/cni-f3ea6fa0-8bd4-7b3c-bcbf-8cc4cdf03de4" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.629 [INFO][4742] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.629 [INFO][4742] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.877 [INFO][4792] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.880 [INFO][4792] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.948 [INFO][4792] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.988 [WARNING][4792] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:53.988 [INFO][4792] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:54.012 [INFO][4792] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:54.110046 containerd[1507]: 2026-03-10 01:31:54.057 [INFO][4742] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:31:54.120746 containerd[1507]: time="2026-03-10T01:31:54.119563132Z" level=info msg="TearDown network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\" successfully" Mar 10 01:31:54.120746 containerd[1507]: time="2026-03-10T01:31:54.119614746Z" level=info msg="StopPodSandbox for \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\" returns successfully" Mar 10 01:31:54.120958 systemd[1]: run-netns-cni\x2df3ea6fa0\x2d8bd4\x2d7b3c\x2dbcbf\x2d8cc4cdf03de4.mount: Deactivated successfully. Mar 10 01:31:54.128447 containerd[1507]: time="2026-03-10T01:31:54.127528830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2d4ds,Uid:ea8ee6c7-d14f-40b8-b33b-78317f32f0c3,Namespace:kube-system,Attempt:1,}" Mar 10 01:31:54.134672 containerd[1507]: time="2026-03-10T01:31:54.128062366Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:54.134672 containerd[1507]: time="2026-03-10T01:31:54.128215414Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:54.134672 containerd[1507]: time="2026-03-10T01:31:54.128406060Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:54.134672 containerd[1507]: time="2026-03-10T01:31:54.128624920Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.132 [INFO][4672] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0 calico-kube-controllers-77f6556c5c- calico-system 2c9450c8-37a4-4965-96d3-67fee0cc59e8 1018 0 2026-03-10 01:31:12 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:77f6556c5c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com calico-kube-controllers-77f6556c5c-68b2f eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calib01fa7c8b48 [] [] }} ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.133 [INFO][4672] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.569 [INFO][4713] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" HandleID="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.634 [INFO][4713] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" HandleID="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f9b0), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"calico-kube-controllers-77f6556c5c-68b2f", "timestamp":"2026-03-10 01:31:53.569140981 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0002ca9a0)} Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.635 [INFO][4713] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.696 [INFO][4713] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.696 [INFO][4713] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.719 [INFO][4713] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.789 [INFO][4713] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.826 [INFO][4713] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.839 [INFO][4713] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.856 [INFO][4713] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.857 [INFO][4713] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.869 [INFO][4713] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1 Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.909 [INFO][4713] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.941 [INFO][4713] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.70/26] block=192.168.57.64/26 handle="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.943 [INFO][4713] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.70/26] handle="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.944 [INFO][4713] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:54.191256 containerd[1507]: 2026-03-10 01:31:53.945 [INFO][4713] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.70/26] IPv6=[] ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" HandleID="k8s-pod-network.c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:54.193417 containerd[1507]: 2026-03-10 01:31:53.981 [INFO][4672] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0", GenerateName:"calico-kube-controllers-77f6556c5c-", Namespace:"calico-system", SelfLink:"", UID:"2c9450c8-37a4-4965-96d3-67fee0cc59e8", ResourceVersion:"1018", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"77f6556c5c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"calico-kube-controllers-77f6556c5c-68b2f", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.57.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib01fa7c8b48", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:54.193417 containerd[1507]: 2026-03-10 01:31:53.984 [INFO][4672] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.70/32] ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:54.193417 containerd[1507]: 2026-03-10 01:31:53.984 [INFO][4672] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib01fa7c8b48 ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:54.193417 containerd[1507]: 2026-03-10 01:31:54.048 [INFO][4672] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:54.193417 containerd[1507]: 2026-03-10 01:31:54.086 [INFO][4672] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0", GenerateName:"calico-kube-controllers-77f6556c5c-", Namespace:"calico-system", SelfLink:"", UID:"2c9450c8-37a4-4965-96d3-67fee0cc59e8", ResourceVersion:"1018", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"77f6556c5c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1", Pod:"calico-kube-controllers-77f6556c5c-68b2f", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.57.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib01fa7c8b48", MAC:"5e:68:a1:9f:24:5d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:54.193417 containerd[1507]: 2026-03-10 01:31:54.142 [INFO][4672] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1" Namespace="calico-system" Pod="calico-kube-controllers-77f6556c5c-68b2f" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:31:54.210520 systemd[1]: Started cri-containerd-b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb.scope - libcontainer container b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb. Mar 10 01:31:54.258687 containerd[1507]: time="2026-03-10T01:31:54.255470948Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-h52bq,Uid:fa9659b6-52d9-4a4d-91d1-6f294264f5ee,Namespace:calico-system,Attempt:1,} returns sandbox id \"50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14\"" Mar 10 01:31:54.257488 systemd[1]: Started cri-containerd-776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020.scope - libcontainer container 776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020. Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:53.805 [INFO][4747] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:53.805 [INFO][4747] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" iface="eth0" netns="/var/run/netns/cni-1ac13a86-5ce3-a999-00d6-f690806454ec" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:53.806 [INFO][4747] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" iface="eth0" netns="/var/run/netns/cni-1ac13a86-5ce3-a999-00d6-f690806454ec" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:53.806 [INFO][4747] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" iface="eth0" netns="/var/run/netns/cni-1ac13a86-5ce3-a999-00d6-f690806454ec" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:53.806 [INFO][4747] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:53.806 [INFO][4747] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:54.250 [INFO][4817] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:54.251 [INFO][4817] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:54.251 [INFO][4817] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:54.288 [WARNING][4817] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:54.288 [INFO][4817] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:54.299 [INFO][4817] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:54.339250 containerd[1507]: 2026-03-10 01:31:54.319 [INFO][4747] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:31:54.341336 containerd[1507]: time="2026-03-10T01:31:54.340993276Z" level=info msg="TearDown network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\" successfully" Mar 10 01:31:54.341336 containerd[1507]: time="2026-03-10T01:31:54.341058792Z" level=info msg="StopPodSandbox for \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\" returns successfully" Mar 10 01:31:54.353280 containerd[1507]: time="2026-03-10T01:31:54.352831849Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-424tn,Uid:c035353a-3dfd-42e9-a291-75ad204d0496,Namespace:calico-system,Attempt:1,}" Mar 10 01:31:54.439013 containerd[1507]: time="2026-03-10T01:31:54.432435421Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:54.439013 containerd[1507]: time="2026-03-10T01:31:54.432553915Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:54.439013 containerd[1507]: time="2026-03-10T01:31:54.432609454Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:54.439013 containerd[1507]: time="2026-03-10T01:31:54.432760363Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:54.564843 containerd[1507]: time="2026-03-10T01:31:54.564762858Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fhcl2,Uid:71ea22a7-e0cb-4bd4-871f-854f4db6a991,Namespace:calico-system,Attempt:1,} returns sandbox id \"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb\"" Mar 10 01:31:54.567724 systemd[1]: Started cri-containerd-c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1.scope - libcontainer container c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1. Mar 10 01:31:54.572686 containerd[1507]: time="2026-03-10T01:31:54.565203060Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-jjvj8,Uid:8c562e47-a863-4e0d-8c19-218349a688f3,Namespace:kube-system,Attempt:1,} returns sandbox id \"776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020\"" Mar 10 01:31:54.641354 containerd[1507]: time="2026-03-10T01:31:54.640719097Z" level=info msg="CreateContainer within sandbox \"776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 10 01:31:54.685515 systemd[1]: run-netns-cni\x2d1ac13a86\x2d5ce3\x2da999\x2d00d6\x2df690806454ec.mount: Deactivated successfully. Mar 10 01:31:54.792260 containerd[1507]: time="2026-03-10T01:31:54.791364260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-77f6556c5c-68b2f,Uid:2c9450c8-37a4-4965-96d3-67fee0cc59e8,Namespace:calico-system,Attempt:1,} returns sandbox id \"c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1\"" Mar 10 01:31:54.802561 systemd-networkd[1419]: cali64cc8014aee: Gained IPv6LL Mar 10 01:31:54.875240 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3572850043.mount: Deactivated successfully. Mar 10 01:31:54.893353 containerd[1507]: time="2026-03-10T01:31:54.891085978Z" level=info msg="CreateContainer within sandbox \"776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b3c176c363251d13a3b6513ad57734e2f1664fe6798ef087beee10254c659874\"" Mar 10 01:31:54.895497 containerd[1507]: time="2026-03-10T01:31:54.895181866Z" level=info msg="StartContainer for \"b3c176c363251d13a3b6513ad57734e2f1664fe6798ef087beee10254c659874\"" Mar 10 01:31:54.897422 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1529194104.mount: Deactivated successfully. Mar 10 01:31:55.007490 systemd-networkd[1419]: calia236e8e9d4d: Link UP Mar 10 01:31:55.014592 systemd-networkd[1419]: calia236e8e9d4d: Gained carrier Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.630 [INFO][4929] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0 coredns-66bc5c9577- kube-system ea8ee6c7-d14f-40b8-b33b-78317f32f0c3 1028 0 2026-03-10 01:30:56 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com coredns-66bc5c9577-2d4ds eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calia236e8e9d4d [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.631 [INFO][4929] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.781 [INFO][5012] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" HandleID="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.828 [INFO][5012] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" HandleID="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fec0), Attrs:map[string]string{"namespace":"kube-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"coredns-66bc5c9577-2d4ds", "timestamp":"2026-03-10 01:31:54.781049915 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc0004b22c0)} Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.831 [INFO][5012] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.831 [INFO][5012] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.831 [INFO][5012] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.839 [INFO][5012] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.854 [INFO][5012] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.868 [INFO][5012] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.879 [INFO][5012] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.889 [INFO][5012] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.889 [INFO][5012] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.900 [INFO][5012] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.909 [INFO][5012] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.930 [INFO][5012] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.71/26] block=192.168.57.64/26 handle="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.930 [INFO][5012] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.71/26] handle="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.931 [INFO][5012] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:55.108414 containerd[1507]: 2026-03-10 01:31:54.931 [INFO][5012] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.71/26] IPv6=[] ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" HandleID="k8s-pod-network.41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:55.111207 containerd[1507]: 2026-03-10 01:31:54.965 [INFO][4929] cni-plugin/k8s.go 418: Populated endpoint ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3", ResourceVersion:"1028", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"coredns-66bc5c9577-2d4ds", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia236e8e9d4d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:55.111207 containerd[1507]: 2026-03-10 01:31:54.978 [INFO][4929] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.71/32] ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:55.111207 containerd[1507]: 2026-03-10 01:31:54.978 [INFO][4929] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia236e8e9d4d ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:55.111207 containerd[1507]: 2026-03-10 01:31:55.031 [INFO][4929] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:55.111207 containerd[1507]: 2026-03-10 01:31:55.050 [INFO][4929] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3", ResourceVersion:"1028", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea", Pod:"coredns-66bc5c9577-2d4ds", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia236e8e9d4d", MAC:"f2:09:c9:15:0c:b7", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:55.113028 containerd[1507]: 2026-03-10 01:31:55.093 [INFO][4929] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea" Namespace="kube-system" Pod="coredns-66bc5c9577-2d4ds" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:31:55.127434 systemd[1]: Started cri-containerd-b3c176c363251d13a3b6513ad57734e2f1664fe6798ef087beee10254c659874.scope - libcontainer container b3c176c363251d13a3b6513ad57734e2f1664fe6798ef087beee10254c659874. Mar 10 01:31:55.140473 systemd-networkd[1419]: calia70ce9298ba: Link UP Mar 10 01:31:55.154535 systemd-networkd[1419]: calia70ce9298ba: Gained carrier Mar 10 01:31:55.184376 systemd-networkd[1419]: calia08e75d8c26: Gained IPv6LL Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.584 [INFO][4953] cni-plugin/plugin.go 342: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0 calico-apiserver-794866df9b- calico-system c035353a-3dfd-42e9-a291-75ad204d0496 1031 0 2026-03-10 01:31:11 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:794866df9b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s srv-eb8wl.gb1.brightbox.com calico-apiserver-794866df9b-424tn eth0 calico-apiserver [] [] [kns.calico-system ksa.calico-system.calico-apiserver] calia70ce9298ba [] [] }} ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.586 [INFO][4953] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.781 [INFO][5006] ipam/ipam_plugin.go 235: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" HandleID="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.828 [INFO][5006] ipam/ipam_plugin.go 301: Auto assigning IP ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" HandleID="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001080d0), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-eb8wl.gb1.brightbox.com", "pod":"calico-apiserver-794866df9b-424tn", "timestamp":"2026-03-10 01:31:54.781640327 +0000 UTC"}, Hostname:"srv-eb8wl.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload", Namespace:(*v1.Namespace)(0xc00041c000)} Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.829 [INFO][5006] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.933 [INFO][5006] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.933 [INFO][5006] ipam/ipam.go 112: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-eb8wl.gb1.brightbox.com' Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.951 [INFO][5006] ipam/ipam.go 707: Looking up existing affinities for host handle="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:54.994 [INFO][5006] ipam/ipam.go 409: Looking up existing affinities for host host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.029 [INFO][5006] ipam/ipam.go 526: Trying affinity for 192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.035 [INFO][5006] ipam/ipam.go 160: Attempting to load block cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.042 [INFO][5006] ipam/ipam.go 237: Affinity is confirmed and block has been loaded cidr=192.168.57.64/26 host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.049 [INFO][5006] ipam/ipam.go 1245: Attempting to assign 1 addresses from block block=192.168.57.64/26 handle="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.053 [INFO][5006] ipam/ipam.go 1806: Creating new handle: k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9 Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.074 [INFO][5006] ipam/ipam.go 1272: Writing block in order to claim IPs block=192.168.57.64/26 handle="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.100 [INFO][5006] ipam/ipam.go 1288: Successfully claimed IPs: [192.168.57.72/26] block=192.168.57.64/26 handle="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.103 [INFO][5006] ipam/ipam.go 895: Auto-assigned 1 out of 1 IPv4s: [192.168.57.72/26] handle="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" host="srv-eb8wl.gb1.brightbox.com" Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.104 [INFO][5006] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:31:55.221413 containerd[1507]: 2026-03-10 01:31:55.104 [INFO][5006] ipam/ipam_plugin.go 325: Calico CNI IPAM assigned addresses IPv4=[192.168.57.72/26] IPv6=[] ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" HandleID="k8s-pod-network.39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:55.224059 containerd[1507]: 2026-03-10 01:31:55.117 [INFO][4953] cni-plugin/k8s.go 418: Populated endpoint ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"c035353a-3dfd-42e9-a291-75ad204d0496", ResourceVersion:"1031", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"", Pod:"calico-apiserver-794866df9b-424tn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calia70ce9298ba", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:55.224059 containerd[1507]: 2026-03-10 01:31:55.120 [INFO][4953] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.57.72/32] ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:55.224059 containerd[1507]: 2026-03-10 01:31:55.121 [INFO][4953] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia70ce9298ba ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:55.224059 containerd[1507]: 2026-03-10 01:31:55.168 [INFO][4953] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:55.224059 containerd[1507]: 2026-03-10 01:31:55.172 [INFO][4953] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"c035353a-3dfd-42e9-a291-75ad204d0496", ResourceVersion:"1031", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9", Pod:"calico-apiserver-794866df9b-424tn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calia70ce9298ba", MAC:"52:24:55:5e:a8:47", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:31:55.224059 containerd[1507]: 2026-03-10 01:31:55.204 [INFO][4953] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9" Namespace="calico-system" Pod="calico-apiserver-794866df9b-424tn" WorkloadEndpoint="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:31:55.248986 systemd-networkd[1419]: calid615bcab3d4: Gained IPv6LL Mar 10 01:31:55.280279 containerd[1507]: time="2026-03-10T01:31:55.275766659Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:55.280279 containerd[1507]: time="2026-03-10T01:31:55.275904998Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:55.280279 containerd[1507]: time="2026-03-10T01:31:55.275931675Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:55.280279 containerd[1507]: time="2026-03-10T01:31:55.276063959Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:55.301074 containerd[1507]: time="2026-03-10T01:31:55.300381280Z" level=info msg="StartContainer for \"b3c176c363251d13a3b6513ad57734e2f1664fe6798ef087beee10254c659874\" returns successfully" Mar 10 01:31:55.376576 systemd[1]: Started cri-containerd-41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea.scope - libcontainer container 41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea. Mar 10 01:31:55.443740 containerd[1507]: time="2026-03-10T01:31:55.440781871Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 10 01:31:55.443740 containerd[1507]: time="2026-03-10T01:31:55.441042969Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 10 01:31:55.443740 containerd[1507]: time="2026-03-10T01:31:55.442500924Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:55.443740 containerd[1507]: time="2026-03-10T01:31:55.442742685Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 10 01:31:55.508887 systemd[1]: Started cri-containerd-39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9.scope - libcontainer container 39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9. Mar 10 01:31:55.574068 containerd[1507]: time="2026-03-10T01:31:55.573995491Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2d4ds,Uid:ea8ee6c7-d14f-40b8-b33b-78317f32f0c3,Namespace:kube-system,Attempt:1,} returns sandbox id \"41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea\"" Mar 10 01:31:55.593910 containerd[1507]: time="2026-03-10T01:31:55.592665291Z" level=info msg="CreateContainer within sandbox \"41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 10 01:31:55.649876 containerd[1507]: time="2026-03-10T01:31:55.649782605Z" level=info msg="CreateContainer within sandbox \"41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"22425bc2934914a2e653a0ba650c2f89b6e7cf126ad514e29603645543131532\"" Mar 10 01:31:55.662034 containerd[1507]: time="2026-03-10T01:31:55.659700434Z" level=info msg="StartContainer for \"22425bc2934914a2e653a0ba650c2f89b6e7cf126ad514e29603645543131532\"" Mar 10 01:31:55.827429 systemd[1]: Started cri-containerd-22425bc2934914a2e653a0ba650c2f89b6e7cf126ad514e29603645543131532.scope - libcontainer container 22425bc2934914a2e653a0ba650c2f89b6e7cf126ad514e29603645543131532. Mar 10 01:31:55.946959 containerd[1507]: time="2026-03-10T01:31:55.946887033Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-794866df9b-424tn,Uid:c035353a-3dfd-42e9-a291-75ad204d0496,Namespace:calico-system,Attempt:1,} returns sandbox id \"39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9\"" Mar 10 01:31:55.997195 containerd[1507]: time="2026-03-10T01:31:55.996687969Z" level=info msg="StartContainer for \"22425bc2934914a2e653a0ba650c2f89b6e7cf126ad514e29603645543131532\" returns successfully" Mar 10 01:31:56.017345 systemd-networkd[1419]: calib01fa7c8b48: Gained IPv6LL Mar 10 01:31:56.184616 kubelet[2683]: I0310 01:31:56.183994 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-jjvj8" podStartSLOduration=60.183913811 podStartE2EDuration="1m0.183913811s" podCreationTimestamp="2026-03-10 01:30:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-10 01:31:56.120348528 +0000 UTC m=+66.227831644" watchObservedRunningTime="2026-03-10 01:31:56.183913811 +0000 UTC m=+66.291396924" Mar 10 01:31:56.276467 kubelet[2683]: I0310 01:31:56.274844 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-2d4ds" podStartSLOduration=60.274820135 podStartE2EDuration="1m0.274820135s" podCreationTimestamp="2026-03-10 01:30:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-10 01:31:56.191256787 +0000 UTC m=+66.298739914" watchObservedRunningTime="2026-03-10 01:31:56.274820135 +0000 UTC m=+66.382303244" Mar 10 01:31:56.913421 systemd-networkd[1419]: calia70ce9298ba: Gained IPv6LL Mar 10 01:31:56.977311 systemd-networkd[1419]: calia236e8e9d4d: Gained IPv6LL Mar 10 01:31:57.352873 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount997385376.mount: Deactivated successfully. Mar 10 01:31:58.254172 containerd[1507]: time="2026-03-10T01:31:58.254103596Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:58.256470 containerd[1507]: time="2026-03-10T01:31:58.256410892Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.31.4: active requests=0, bytes read=55623386" Mar 10 01:31:58.257645 containerd[1507]: time="2026-03-10T01:31:58.257566659Z" level=info msg="ImageCreate event name:\"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:58.263539 containerd[1507]: time="2026-03-10T01:31:58.261106720Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:31:58.263539 containerd[1507]: time="2026-03-10T01:31:58.262517595Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" with image id \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:44395ca5ebfe88f21ed51acfbec5fc0f31d2762966e2007a0a2eb9b30e35fc4d\", size \"55623232\" in 7.022858332s" Mar 10 01:31:58.263539 containerd[1507]: time="2026-03-10T01:31:58.262563905Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.31.4\" returns image reference \"sha256:714983e5e920bbe810fab04d9f06bd16ef4e552b0d2deffd7ab2b2c4a001acbb\"" Mar 10 01:31:58.274620 containerd[1507]: time="2026-03-10T01:31:58.274581238Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Mar 10 01:31:58.283125 containerd[1507]: time="2026-03-10T01:31:58.282746915Z" level=info msg="CreateContainer within sandbox \"c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Mar 10 01:31:58.318309 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3565276428.mount: Deactivated successfully. Mar 10 01:31:58.350061 containerd[1507]: time="2026-03-10T01:31:58.349845986Z" level=info msg="CreateContainer within sandbox \"c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93\"" Mar 10 01:31:58.350984 containerd[1507]: time="2026-03-10T01:31:58.350932129Z" level=info msg="StartContainer for \"2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93\"" Mar 10 01:31:58.427748 systemd[1]: run-containerd-runc-k8s.io-2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93-runc.L9iq4V.mount: Deactivated successfully. Mar 10 01:31:58.439448 systemd[1]: Started cri-containerd-2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93.scope - libcontainer container 2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93. Mar 10 01:31:58.519007 containerd[1507]: time="2026-03-10T01:31:58.518521460Z" level=info msg="StartContainer for \"2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93\" returns successfully" Mar 10 01:31:59.209594 kubelet[2683]: I0310 01:31:59.207099 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-cccfbd5cf-sxw9t" podStartSLOduration=41.170179841 podStartE2EDuration="48.207063605s" podCreationTimestamp="2026-03-10 01:31:11 +0000 UTC" firstStartedPulling="2026-03-10 01:31:51.237029272 +0000 UTC m=+61.344512380" lastFinishedPulling="2026-03-10 01:31:58.273913035 +0000 UTC m=+68.381396144" observedRunningTime="2026-03-10 01:31:59.204301543 +0000 UTC m=+69.311784673" watchObservedRunningTime="2026-03-10 01:31:59.207063605 +0000 UTC m=+69.314546715" Mar 10 01:32:00.277656 systemd[1]: run-containerd-runc-k8s.io-2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93-runc.0Nf99W.mount: Deactivated successfully. Mar 10 01:32:01.312912 systemd[1]: run-containerd-runc-k8s.io-2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93-runc.krkchN.mount: Deactivated successfully. Mar 10 01:32:02.338639 containerd[1507]: time="2026-03-10T01:32:02.338544938Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:02.340307 containerd[1507]: time="2026-03-10T01:32:02.340187689Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=48415780" Mar 10 01:32:02.341179 containerd[1507]: time="2026-03-10T01:32:02.341101919Z" level=info msg="ImageCreate event name:\"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:02.353551 containerd[1507]: time="2026-03-10T01:32:02.353463069Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:02.355248 containerd[1507]: time="2026-03-10T01:32:02.354938326Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 4.080300798s" Mar 10 01:32:02.355248 containerd[1507]: time="2026-03-10T01:32:02.355001862Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Mar 10 01:32:02.357366 containerd[1507]: time="2026-03-10T01:32:02.357206154Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\"" Mar 10 01:32:02.365347 containerd[1507]: time="2026-03-10T01:32:02.365024397Z" level=info msg="CreateContainer within sandbox \"50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 10 01:32:02.385343 containerd[1507]: time="2026-03-10T01:32:02.385180492Z" level=info msg="CreateContainer within sandbox \"50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"d452d8200704e2db625ba72c9fa26f1f98e60964b2647f8913708db015d4e999\"" Mar 10 01:32:02.386499 containerd[1507]: time="2026-03-10T01:32:02.386427188Z" level=info msg="StartContainer for \"d452d8200704e2db625ba72c9fa26f1f98e60964b2647f8913708db015d4e999\"" Mar 10 01:32:02.466438 systemd[1]: Started cri-containerd-d452d8200704e2db625ba72c9fa26f1f98e60964b2647f8913708db015d4e999.scope - libcontainer container d452d8200704e2db625ba72c9fa26f1f98e60964b2647f8913708db015d4e999. Mar 10 01:32:02.537446 containerd[1507]: time="2026-03-10T01:32:02.537364215Z" level=info msg="StartContainer for \"d452d8200704e2db625ba72c9fa26f1f98e60964b2647f8913708db015d4e999\" returns successfully" Mar 10 01:32:03.415874 kubelet[2683]: I0310 01:32:03.408954 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-794866df9b-h52bq" podStartSLOduration=44.310765719 podStartE2EDuration="52.391958235s" podCreationTimestamp="2026-03-10 01:31:11 +0000 UTC" firstStartedPulling="2026-03-10 01:31:54.275735297 +0000 UTC m=+64.383218401" lastFinishedPulling="2026-03-10 01:32:02.35692781 +0000 UTC m=+72.464410917" observedRunningTime="2026-03-10 01:32:03.389619249 +0000 UTC m=+73.497102362" watchObservedRunningTime="2026-03-10 01:32:03.391958235 +0000 UTC m=+73.499441355" Mar 10 01:32:04.332413 containerd[1507]: time="2026-03-10T01:32:04.332319252Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:04.335110 containerd[1507]: time="2026-03-10T01:32:04.333356048Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.31.4: active requests=0, bytes read=8792502" Mar 10 01:32:04.339580 containerd[1507]: time="2026-03-10T01:32:04.338393128Z" level=info msg="ImageCreate event name:\"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:04.345387 containerd[1507]: time="2026-03-10T01:32:04.345291732Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:04.347031 containerd[1507]: time="2026-03-10T01:32:04.346987885Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.31.4\" with image id \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\", repo tag \"ghcr.io/flatcar/calico/csi:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:ab57dd6f8423ef7b3ff382bf4ca5ace6063bdca77d441d852c75ec58847dd280\", size \"10348547\" in 1.989695401s" Mar 10 01:32:04.347164 containerd[1507]: time="2026-03-10T01:32:04.347035869Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.31.4\" returns image reference \"sha256:4c8cd7d0b10a4df64a5bd90e9845e9d1edbe0e37c2ebfc171bb28698e07abf72\"" Mar 10 01:32:04.353539 kubelet[2683]: I0310 01:32:04.353492 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 10 01:32:04.357294 containerd[1507]: time="2026-03-10T01:32:04.355091639Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\"" Mar 10 01:32:04.375592 containerd[1507]: time="2026-03-10T01:32:04.375002685Z" level=info msg="CreateContainer within sandbox \"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 10 01:32:04.426533 containerd[1507]: time="2026-03-10T01:32:04.426328230Z" level=info msg="CreateContainer within sandbox \"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"d1053de3e02a0600d85adff31d80b7d1cc09ad0c2e9c325336e356b6144cec61\"" Mar 10 01:32:04.429287 containerd[1507]: time="2026-03-10T01:32:04.427904011Z" level=info msg="StartContainer for \"d1053de3e02a0600d85adff31d80b7d1cc09ad0c2e9c325336e356b6144cec61\"" Mar 10 01:32:04.503425 systemd[1]: Started cri-containerd-d1053de3e02a0600d85adff31d80b7d1cc09ad0c2e9c325336e356b6144cec61.scope - libcontainer container d1053de3e02a0600d85adff31d80b7d1cc09ad0c2e9c325336e356b6144cec61. Mar 10 01:32:04.576794 containerd[1507]: time="2026-03-10T01:32:04.576731460Z" level=info msg="StartContainer for \"d1053de3e02a0600d85adff31d80b7d1cc09ad0c2e9c325336e356b6144cec61\" returns successfully" Mar 10 01:32:12.271894 containerd[1507]: time="2026-03-10T01:32:12.271447816Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:12.275925 containerd[1507]: time="2026-03-10T01:32:12.275825446Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.31.4: active requests=0, bytes read=52406348" Mar 10 01:32:12.276128 containerd[1507]: time="2026-03-10T01:32:12.276069996Z" level=info msg="ImageCreate event name:\"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:12.287254 containerd[1507]: time="2026-03-10T01:32:12.286436003Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:12.287924 containerd[1507]: time="2026-03-10T01:32:12.287883439Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" with image id \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:99b8bb50141ca55b4b6ddfcf2f2fbde838265508ab2ac96ed08e72cd39800713\", size \"53962361\" in 7.932745415s" Mar 10 01:32:12.288004 containerd[1507]: time="2026-03-10T01:32:12.287945229Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.31.4\" returns image reference \"sha256:ff033cc89dab51090bfa1b04e155a5ce1e3b1f324f74b7b2be0dd6f0b6b10e89\"" Mar 10 01:32:12.301202 containerd[1507]: time="2026-03-10T01:32:12.301141452Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\"" Mar 10 01:32:12.434717 containerd[1507]: time="2026-03-10T01:32:12.434648501Z" level=info msg="CreateContainer within sandbox \"c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 10 01:32:12.534829 systemd[1]: Started sshd@9-10.230.50.50:22-68.220.241.50:44514.service - OpenSSH per-connection server daemon (68.220.241.50:44514). Mar 10 01:32:12.680863 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount835723054.mount: Deactivated successfully. Mar 10 01:32:12.791760 containerd[1507]: time="2026-03-10T01:32:12.791577856Z" level=info msg="CreateContainer within sandbox \"c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"f57723e5eba202cab16769af7356cebb41cf340572d36142435bb5da4f771080\"" Mar 10 01:32:12.798287 containerd[1507]: time="2026-03-10T01:32:12.797785771Z" level=info msg="StartContainer for \"f57723e5eba202cab16769af7356cebb41cf340572d36142435bb5da4f771080\"" Mar 10 01:32:12.973259 containerd[1507]: time="2026-03-10T01:32:12.969465037Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:12.985773 containerd[1507]: time="2026-03-10T01:32:12.983516546Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.31.4: active requests=0, bytes read=77" Mar 10 01:32:13.000978 containerd[1507]: time="2026-03-10T01:32:13.000904300Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" with image id \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:d212af1da3dd52a633bc9e36653a7d901d95a570f8d51d1968a837dcf6879730\", size \"49971841\" in 699.688488ms" Mar 10 01:32:13.001351 containerd[1507]: time="2026-03-10T01:32:13.001150618Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.31.4\" returns image reference \"sha256:f7ff80340b9b4973ceda29859065985831588b2898f2b4009f742b5789010898\"" Mar 10 01:32:13.007885 containerd[1507]: time="2026-03-10T01:32:13.007366720Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\"" Mar 10 01:32:13.024640 containerd[1507]: time="2026-03-10T01:32:13.024496146Z" level=info msg="CreateContainer within sandbox \"39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 10 01:32:13.087584 containerd[1507]: time="2026-03-10T01:32:13.087541746Z" level=info msg="CreateContainer within sandbox \"39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"0e725a93a931d7b75aea92b3daa28d99c04bf0d4fa08c9be438159c0a0146154\"" Mar 10 01:32:13.091395 containerd[1507]: time="2026-03-10T01:32:13.091363428Z" level=info msg="StartContainer for \"0e725a93a931d7b75aea92b3daa28d99c04bf0d4fa08c9be438159c0a0146154\"" Mar 10 01:32:13.291417 systemd[1]: Started cri-containerd-f57723e5eba202cab16769af7356cebb41cf340572d36142435bb5da4f771080.scope - libcontainer container f57723e5eba202cab16769af7356cebb41cf340572d36142435bb5da4f771080. Mar 10 01:32:13.352858 systemd[1]: Started cri-containerd-0e725a93a931d7b75aea92b3daa28d99c04bf0d4fa08c9be438159c0a0146154.scope - libcontainer container 0e725a93a931d7b75aea92b3daa28d99c04bf0d4fa08c9be438159c0a0146154. Mar 10 01:32:13.375616 sshd[5495]: Accepted publickey for core from 68.220.241.50 port 44514 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:13.387951 sshd[5495]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:13.433751 systemd-logind[1485]: New session 12 of user core. Mar 10 01:32:13.438551 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 10 01:32:13.672760 containerd[1507]: time="2026-03-10T01:32:13.672538814Z" level=info msg="StartContainer for \"0e725a93a931d7b75aea92b3daa28d99c04bf0d4fa08c9be438159c0a0146154\" returns successfully" Mar 10 01:32:13.683633 containerd[1507]: time="2026-03-10T01:32:13.683440342Z" level=info msg="StartContainer for \"f57723e5eba202cab16769af7356cebb41cf340572d36142435bb5da4f771080\" returns successfully" Mar 10 01:32:14.860869 sshd[5495]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:14.892865 systemd-logind[1485]: Session 12 logged out. Waiting for processes to exit. Mar 10 01:32:14.896673 systemd[1]: sshd@9-10.230.50.50:22-68.220.241.50:44514.service: Deactivated successfully. Mar 10 01:32:14.900627 systemd[1]: session-12.scope: Deactivated successfully. Mar 10 01:32:14.903698 systemd-logind[1485]: Removed session 12. Mar 10 01:32:14.942702 kubelet[2683]: I0310 01:32:14.937302 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-77f6556c5c-68b2f" podStartSLOduration=45.430960541 podStartE2EDuration="1m2.918314472s" podCreationTimestamp="2026-03-10 01:31:12 +0000 UTC" firstStartedPulling="2026-03-10 01:31:54.813448408 +0000 UTC m=+64.920931514" lastFinishedPulling="2026-03-10 01:32:12.300802336 +0000 UTC m=+82.408285445" observedRunningTime="2026-03-10 01:32:14.860182379 +0000 UTC m=+84.967665498" watchObservedRunningTime="2026-03-10 01:32:14.918314472 +0000 UTC m=+85.025797594" Mar 10 01:32:15.058419 kubelet[2683]: I0310 01:32:15.057404 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-apiserver-794866df9b-424tn" podStartSLOduration=47.00717789 podStartE2EDuration="1m4.057378476s" podCreationTimestamp="2026-03-10 01:31:11 +0000 UTC" firstStartedPulling="2026-03-10 01:31:55.954151785 +0000 UTC m=+66.061634892" lastFinishedPulling="2026-03-10 01:32:13.004352372 +0000 UTC m=+83.111835478" observedRunningTime="2026-03-10 01:32:14.944273372 +0000 UTC m=+85.051756508" watchObservedRunningTime="2026-03-10 01:32:15.057378476 +0000 UTC m=+85.164861589" Mar 10 01:32:16.749380 kubelet[2683]: I0310 01:32:16.749028 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 10 01:32:18.088479 containerd[1507]: time="2026-03-10T01:32:18.088244523Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:18.090664 containerd[1507]: time="2026-03-10T01:32:18.090051924Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4: active requests=0, bytes read=14704317" Mar 10 01:32:18.090664 containerd[1507]: time="2026-03-10T01:32:18.090266933Z" level=info msg="ImageCreate event name:\"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:18.106173 containerd[1507]: time="2026-03-10T01:32:18.105916656Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 10 01:32:18.107480 containerd[1507]: time="2026-03-10T01:32:18.107213747Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" with image id \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:e41c0d73bcd33ff28ae2f2983cf781a4509d212e102d53883dbbf436ab3cd97d\", size \"16260314\" in 5.099786364s" Mar 10 01:32:18.107480 containerd[1507]: time="2026-03-10T01:32:18.107286477Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.31.4\" returns image reference \"sha256:d7aeb99114cbb6499e9048f43d3faa5f199d1a05ed44165e5974d0368ac32771\"" Mar 10 01:32:18.233234 containerd[1507]: time="2026-03-10T01:32:18.233126948Z" level=info msg="CreateContainer within sandbox \"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 10 01:32:18.266328 containerd[1507]: time="2026-03-10T01:32:18.262411870Z" level=info msg="CreateContainer within sandbox \"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"96989c5e97b384ddcf3072cf1b9b3b5de7256387775e6a1fc5dcaa84fca219ae\"" Mar 10 01:32:18.267619 containerd[1507]: time="2026-03-10T01:32:18.266416261Z" level=info msg="StartContainer for \"96989c5e97b384ddcf3072cf1b9b3b5de7256387775e6a1fc5dcaa84fca219ae\"" Mar 10 01:32:18.401723 systemd[1]: run-containerd-runc-k8s.io-96989c5e97b384ddcf3072cf1b9b3b5de7256387775e6a1fc5dcaa84fca219ae-runc.TxGxMM.mount: Deactivated successfully. Mar 10 01:32:18.421179 systemd[1]: Started cri-containerd-96989c5e97b384ddcf3072cf1b9b3b5de7256387775e6a1fc5dcaa84fca219ae.scope - libcontainer container 96989c5e97b384ddcf3072cf1b9b3b5de7256387775e6a1fc5dcaa84fca219ae. Mar 10 01:32:18.480223 containerd[1507]: time="2026-03-10T01:32:18.480163281Z" level=info msg="StartContainer for \"96989c5e97b384ddcf3072cf1b9b3b5de7256387775e6a1fc5dcaa84fca219ae\" returns successfully" Mar 10 01:32:18.666497 kubelet[2683]: I0310 01:32:18.663918 2683 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 10 01:32:18.669120 kubelet[2683]: I0310 01:32:18.668071 2683 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 10 01:32:19.983343 systemd[1]: Started sshd@10-10.230.50.50:22-68.220.241.50:44522.service - OpenSSH per-connection server daemon (68.220.241.50:44522). Mar 10 01:32:20.652345 sshd[5676]: Accepted publickey for core from 68.220.241.50 port 44522 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:20.675908 sshd[5676]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:20.687662 systemd-logind[1485]: New session 13 of user core. Mar 10 01:32:20.695347 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 10 01:32:21.621131 sshd[5676]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:21.629134 systemd[1]: sshd@10-10.230.50.50:22-68.220.241.50:44522.service: Deactivated successfully. Mar 10 01:32:21.633320 systemd[1]: session-13.scope: Deactivated successfully. Mar 10 01:32:21.636242 systemd-logind[1485]: Session 13 logged out. Waiting for processes to exit. Mar 10 01:32:21.638533 systemd-logind[1485]: Removed session 13. Mar 10 01:32:23.769724 systemd[1]: run-containerd-runc-k8s.io-2bae45b5466b634fb3dc8696e3ea762f9f86d888180715731392a1940c82fb93-runc.QRjy9m.mount: Deactivated successfully. Mar 10 01:32:26.738588 systemd[1]: Started sshd@11-10.230.50.50:22-68.220.241.50:38944.service - OpenSSH per-connection server daemon (68.220.241.50:38944). Mar 10 01:32:27.401962 sshd[5742]: Accepted publickey for core from 68.220.241.50 port 38944 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:27.405574 sshd[5742]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:27.414947 systemd-logind[1485]: New session 14 of user core. Mar 10 01:32:27.421422 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 10 01:32:28.323945 sshd[5742]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:28.329894 systemd[1]: sshd@11-10.230.50.50:22-68.220.241.50:38944.service: Deactivated successfully. Mar 10 01:32:28.332955 systemd[1]: session-14.scope: Deactivated successfully. Mar 10 01:32:28.335182 systemd-logind[1485]: Session 14 logged out. Waiting for processes to exit. Mar 10 01:32:28.337287 systemd-logind[1485]: Removed session 14. Mar 10 01:32:33.436481 systemd[1]: Started sshd@12-10.230.50.50:22-68.220.241.50:54324.service - OpenSSH per-connection server daemon (68.220.241.50:54324). Mar 10 01:32:34.097940 sshd[5776]: Accepted publickey for core from 68.220.241.50 port 54324 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:34.100929 sshd[5776]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:34.110435 systemd-logind[1485]: New session 15 of user core. Mar 10 01:32:34.120499 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 10 01:32:34.694740 sshd[5776]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:34.703536 systemd[1]: sshd@12-10.230.50.50:22-68.220.241.50:54324.service: Deactivated successfully. Mar 10 01:32:34.707112 systemd[1]: session-15.scope: Deactivated successfully. Mar 10 01:32:34.709058 systemd-logind[1485]: Session 15 logged out. Waiting for processes to exit. Mar 10 01:32:34.711199 systemd-logind[1485]: Removed session 15. Mar 10 01:32:39.807783 systemd[1]: Started sshd@13-10.230.50.50:22-68.220.241.50:54336.service - OpenSSH per-connection server daemon (68.220.241.50:54336). Mar 10 01:32:40.445941 sshd[5796]: Accepted publickey for core from 68.220.241.50 port 54336 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:40.456941 sshd[5796]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:40.474082 systemd-logind[1485]: New session 16 of user core. Mar 10 01:32:40.480981 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 10 01:32:41.360823 sshd[5796]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:41.371907 systemd[1]: sshd@13-10.230.50.50:22-68.220.241.50:54336.service: Deactivated successfully. Mar 10 01:32:41.377725 systemd[1]: session-16.scope: Deactivated successfully. Mar 10 01:32:41.379238 systemd-logind[1485]: Session 16 logged out. Waiting for processes to exit. Mar 10 01:32:41.381068 systemd-logind[1485]: Removed session 16. Mar 10 01:32:41.465999 systemd[1]: Started sshd@14-10.230.50.50:22-68.220.241.50:54340.service - OpenSSH per-connection server daemon (68.220.241.50:54340). Mar 10 01:32:42.102445 sshd[5849]: Accepted publickey for core from 68.220.241.50 port 54340 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:42.107904 sshd[5849]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:42.123568 systemd-logind[1485]: New session 17 of user core. Mar 10 01:32:42.132584 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 10 01:32:42.732160 sshd[5849]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:42.748795 systemd[1]: sshd@14-10.230.50.50:22-68.220.241.50:54340.service: Deactivated successfully. Mar 10 01:32:42.754419 systemd[1]: session-17.scope: Deactivated successfully. Mar 10 01:32:42.759701 systemd-logind[1485]: Session 17 logged out. Waiting for processes to exit. Mar 10 01:32:42.762955 systemd-logind[1485]: Removed session 17. Mar 10 01:32:42.835599 systemd[1]: Started sshd@15-10.230.50.50:22-68.220.241.50:55796.service - OpenSSH per-connection server daemon (68.220.241.50:55796). Mar 10 01:32:43.466679 sshd[5860]: Accepted publickey for core from 68.220.241.50 port 55796 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:43.469966 sshd[5860]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:43.477840 systemd-logind[1485]: New session 18 of user core. Mar 10 01:32:43.483461 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 10 01:32:44.176109 sshd[5860]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:44.186746 systemd[1]: sshd@15-10.230.50.50:22-68.220.241.50:55796.service: Deactivated successfully. Mar 10 01:32:44.190581 systemd[1]: session-18.scope: Deactivated successfully. Mar 10 01:32:44.192597 systemd-logind[1485]: Session 18 logged out. Waiting for processes to exit. Mar 10 01:32:44.194246 systemd-logind[1485]: Removed session 18. Mar 10 01:32:49.283645 systemd[1]: Started sshd@16-10.230.50.50:22-68.220.241.50:55798.service - OpenSSH per-connection server daemon (68.220.241.50:55798). Mar 10 01:32:49.867457 sshd[5892]: Accepted publickey for core from 68.220.241.50 port 55798 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:49.870258 sshd[5892]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:49.882140 systemd-logind[1485]: New session 19 of user core. Mar 10 01:32:49.889499 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 10 01:32:50.629954 sshd[5892]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:50.643589 systemd[1]: sshd@16-10.230.50.50:22-68.220.241.50:55798.service: Deactivated successfully. Mar 10 01:32:50.647698 systemd[1]: session-19.scope: Deactivated successfully. Mar 10 01:32:50.649351 systemd-logind[1485]: Session 19 logged out. Waiting for processes to exit. Mar 10 01:32:50.650828 systemd-logind[1485]: Removed session 19. Mar 10 01:32:51.012019 containerd[1507]: time="2026-03-10T01:32:51.004354751Z" level=info msg="StopPodSandbox for \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\"" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.385 [WARNING][5914] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"71ea22a7-e0cb-4bd4-871f-854f4db6a991", ResourceVersion:"1231", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"98cbb5577", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb", Pod:"csi-node-driver-fhcl2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.57.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia08e75d8c26", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.387 [INFO][5914] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.387 [INFO][5914] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" iface="eth0" netns="" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.388 [INFO][5914] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.388 [INFO][5914] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.600 [INFO][5921] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.603 [INFO][5921] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.603 [INFO][5921] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.619 [WARNING][5921] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.622 [INFO][5921] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.625 [INFO][5921] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:51.631276 containerd[1507]: 2026-03-10 01:32:51.628 [INFO][5914] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.639578 containerd[1507]: time="2026-03-10T01:32:51.639520797Z" level=info msg="TearDown network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\" successfully" Mar 10 01:32:51.639659 containerd[1507]: time="2026-03-10T01:32:51.639597359Z" level=info msg="StopPodSandbox for \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\" returns successfully" Mar 10 01:32:51.656351 containerd[1507]: time="2026-03-10T01:32:51.655895590Z" level=info msg="RemovePodSandbox for \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\"" Mar 10 01:32:51.664487 containerd[1507]: time="2026-03-10T01:32:51.664404322Z" level=info msg="Forcibly stopping sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\"" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.732 [WARNING][5936] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"71ea22a7-e0cb-4bd4-871f-854f4db6a991", ResourceVersion:"1231", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"98cbb5577", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"b0a8c5fd68fc493ad795516c7cd230dbaaae379d8528cb899ca48df308f1b4fb", Pod:"csi-node-driver-fhcl2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.57.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia08e75d8c26", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.732 [INFO][5936] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.732 [INFO][5936] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" iface="eth0" netns="" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.732 [INFO][5936] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.732 [INFO][5936] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.770 [INFO][5943] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.771 [INFO][5943] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.771 [INFO][5943] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.781 [WARNING][5943] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.781 [INFO][5943] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" HandleID="k8s-pod-network.21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Workload="srv--eb8wl.gb1.brightbox.com-k8s-csi--node--driver--fhcl2-eth0" Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.783 [INFO][5943] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:51.792208 containerd[1507]: 2026-03-10 01:32:51.787 [INFO][5936] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f" Mar 10 01:32:51.792208 containerd[1507]: time="2026-03-10T01:32:51.791868233Z" level=info msg="TearDown network for sandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\" successfully" Mar 10 01:32:51.834418 containerd[1507]: time="2026-03-10T01:32:51.834307362Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:32:51.834878 containerd[1507]: time="2026-03-10T01:32:51.834483999Z" level=info msg="RemovePodSandbox \"21d5822b1ef4b7d0ffc4501dc708dd8b21432742eb9c6e7158f0141680b4090f\" returns successfully" Mar 10 01:32:51.836550 containerd[1507]: time="2026-03-10T01:32:51.836484619Z" level=info msg="StopPodSandbox for \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\"" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.900 [WARNING][5957] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0", GenerateName:"calico-kube-controllers-77f6556c5c-", Namespace:"calico-system", SelfLink:"", UID:"2c9450c8-37a4-4965-96d3-67fee0cc59e8", ResourceVersion:"1199", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"77f6556c5c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1", Pod:"calico-kube-controllers-77f6556c5c-68b2f", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.57.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib01fa7c8b48", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.901 [INFO][5957] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.901 [INFO][5957] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" iface="eth0" netns="" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.901 [INFO][5957] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.901 [INFO][5957] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.944 [INFO][5964] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.944 [INFO][5964] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.944 [INFO][5964] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.954 [WARNING][5964] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.954 [INFO][5964] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.956 [INFO][5964] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:51.962927 containerd[1507]: 2026-03-10 01:32:51.959 [INFO][5957] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:51.962927 containerd[1507]: time="2026-03-10T01:32:51.962598801Z" level=info msg="TearDown network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\" successfully" Mar 10 01:32:51.962927 containerd[1507]: time="2026-03-10T01:32:51.962638796Z" level=info msg="StopPodSandbox for \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\" returns successfully" Mar 10 01:32:51.967051 containerd[1507]: time="2026-03-10T01:32:51.964709945Z" level=info msg="RemovePodSandbox for \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\"" Mar 10 01:32:51.967051 containerd[1507]: time="2026-03-10T01:32:51.964747305Z" level=info msg="Forcibly stopping sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\"" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.026 [WARNING][5978] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0", GenerateName:"calico-kube-controllers-77f6556c5c-", Namespace:"calico-system", SelfLink:"", UID:"2c9450c8-37a4-4965-96d3-67fee0cc59e8", ResourceVersion:"1199", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"77f6556c5c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"c6ad37b9afabd1d63bb8fcc1ca15ac0f586c8e2e68516d30385b56a5117a92d1", Pod:"calico-kube-controllers-77f6556c5c-68b2f", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.57.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib01fa7c8b48", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.026 [INFO][5978] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.026 [INFO][5978] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" iface="eth0" netns="" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.026 [INFO][5978] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.026 [INFO][5978] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.073 [INFO][5985] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.073 [INFO][5985] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.074 [INFO][5985] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.085 [WARNING][5985] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.085 [INFO][5985] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" HandleID="k8s-pod-network.e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--kube--controllers--77f6556c5c--68b2f-eth0" Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.087 [INFO][5985] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:52.093342 containerd[1507]: 2026-03-10 01:32:52.090 [INFO][5978] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1" Mar 10 01:32:52.095321 containerd[1507]: time="2026-03-10T01:32:52.093507127Z" level=info msg="TearDown network for sandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\" successfully" Mar 10 01:32:52.144382 containerd[1507]: time="2026-03-10T01:32:52.144305574Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:32:52.145354 containerd[1507]: time="2026-03-10T01:32:52.144407468Z" level=info msg="RemovePodSandbox \"e3b2f61fde9a41fba75b68f95d5d840c619d3f6fc6246ad7979562f462eb6af1\" returns successfully" Mar 10 01:32:52.145354 containerd[1507]: time="2026-03-10T01:32:52.145170466Z" level=info msg="StopPodSandbox for \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\"" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.214 [WARNING][5999] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"8c562e47-a863-4e0d-8c19-218349a688f3", ResourceVersion:"1064", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020", Pod:"coredns-66bc5c9577-jjvj8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid615bcab3d4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.215 [INFO][5999] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.215 [INFO][5999] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" iface="eth0" netns="" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.215 [INFO][5999] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.215 [INFO][5999] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.299 [INFO][6006] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.302 [INFO][6006] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.302 [INFO][6006] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.316 [WARNING][6006] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.316 [INFO][6006] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.319 [INFO][6006] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:52.325938 containerd[1507]: 2026-03-10 01:32:52.322 [INFO][5999] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.325938 containerd[1507]: time="2026-03-10T01:32:52.325381262Z" level=info msg="TearDown network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\" successfully" Mar 10 01:32:52.325938 containerd[1507]: time="2026-03-10T01:32:52.325427668Z" level=info msg="StopPodSandbox for \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\" returns successfully" Mar 10 01:32:52.330748 containerd[1507]: time="2026-03-10T01:32:52.329011244Z" level=info msg="RemovePodSandbox for \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\"" Mar 10 01:32:52.330748 containerd[1507]: time="2026-03-10T01:32:52.329052323Z" level=info msg="Forcibly stopping sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\"" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.392 [WARNING][6021] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"8c562e47-a863-4e0d-8c19-218349a688f3", ResourceVersion:"1064", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"776aaabfdb8c6563d2b590eb215420faf5affe4f2f66e2d3ed61cc75a64f2020", Pod:"coredns-66bc5c9577-jjvj8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid615bcab3d4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.393 [INFO][6021] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.393 [INFO][6021] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" iface="eth0" netns="" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.393 [INFO][6021] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.393 [INFO][6021] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.436 [INFO][6028] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.436 [INFO][6028] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.436 [INFO][6028] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.448 [WARNING][6028] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.448 [INFO][6028] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" HandleID="k8s-pod-network.1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--jjvj8-eth0" Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.450 [INFO][6028] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:52.456537 containerd[1507]: 2026-03-10 01:32:52.453 [INFO][6021] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3" Mar 10 01:32:52.459262 containerd[1507]: time="2026-03-10T01:32:52.457603353Z" level=info msg="TearDown network for sandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\" successfully" Mar 10 01:32:52.462109 containerd[1507]: time="2026-03-10T01:32:52.462044447Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:32:52.462396 containerd[1507]: time="2026-03-10T01:32:52.462266353Z" level=info msg="RemovePodSandbox \"1a8e39adef8f45c5ba0f541dc664ee4aabea7ec0d5629e3a0bb039ed023cb7f3\" returns successfully" Mar 10 01:32:52.463384 containerd[1507]: time="2026-03-10T01:32:52.463293029Z" level=info msg="StopPodSandbox for \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\"" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.527 [WARNING][6043] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"c035353a-3dfd-42e9-a291-75ad204d0496", ResourceVersion:"1209", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9", Pod:"calico-apiserver-794866df9b-424tn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calia70ce9298ba", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.527 [INFO][6043] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.527 [INFO][6043] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" iface="eth0" netns="" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.527 [INFO][6043] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.527 [INFO][6043] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.565 [INFO][6051] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.565 [INFO][6051] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.566 [INFO][6051] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.577 [WARNING][6051] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.577 [INFO][6051] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.579 [INFO][6051] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:52.586410 containerd[1507]: 2026-03-10 01:32:52.583 [INFO][6043] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.589350 containerd[1507]: time="2026-03-10T01:32:52.586466798Z" level=info msg="TearDown network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\" successfully" Mar 10 01:32:52.589350 containerd[1507]: time="2026-03-10T01:32:52.586501614Z" level=info msg="StopPodSandbox for \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\" returns successfully" Mar 10 01:32:52.589350 containerd[1507]: time="2026-03-10T01:32:52.587503866Z" level=info msg="RemovePodSandbox for \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\"" Mar 10 01:32:52.589350 containerd[1507]: time="2026-03-10T01:32:52.587543118Z" level=info msg="Forcibly stopping sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\"" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.643 [WARNING][6066] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"c035353a-3dfd-42e9-a291-75ad204d0496", ResourceVersion:"1209", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"39456c685c3c0b968b1e44a24406bbe09d2501c4057810067b8afb8983fb81b9", Pod:"calico-apiserver-794866df9b-424tn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"calia70ce9298ba", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.644 [INFO][6066] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.644 [INFO][6066] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" iface="eth0" netns="" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.644 [INFO][6066] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.644 [INFO][6066] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.683 [INFO][6074] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.683 [INFO][6074] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.683 [INFO][6074] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.693 [WARNING][6074] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.693 [INFO][6074] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" HandleID="k8s-pod-network.e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--424tn-eth0" Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.695 [INFO][6074] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:52.701273 containerd[1507]: 2026-03-10 01:32:52.698 [INFO][6066] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0" Mar 10 01:32:52.703181 containerd[1507]: time="2026-03-10T01:32:52.701476852Z" level=info msg="TearDown network for sandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\" successfully" Mar 10 01:32:52.706076 containerd[1507]: time="2026-03-10T01:32:52.706036775Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:32:52.706209 containerd[1507]: time="2026-03-10T01:32:52.706115900Z" level=info msg="RemovePodSandbox \"e2b6be2471293579a809812f6822e1538487104cc0846299b23a6d7e5816edb0\" returns successfully" Mar 10 01:32:52.707287 containerd[1507]: time="2026-03-10T01:32:52.707012948Z" level=info msg="StopPodSandbox for \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\"" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.777 [WARNING][6088] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0", GenerateName:"goldmane-cccfbd5cf-", Namespace:"calico-system", SelfLink:"", UID:"4c86706d-5198-4966-aa4f-480259ec456a", ResourceVersion:"1091", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"cccfbd5cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42", Pod:"goldmane-cccfbd5cf-sxw9t", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.57.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calid991037db6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.777 [INFO][6088] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.777 [INFO][6088] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" iface="eth0" netns="" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.777 [INFO][6088] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.777 [INFO][6088] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.821 [INFO][6095] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.821 [INFO][6095] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.821 [INFO][6095] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.833 [WARNING][6095] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.833 [INFO][6095] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.835 [INFO][6095] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:52.841355 containerd[1507]: 2026-03-10 01:32:52.838 [INFO][6088] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.841355 containerd[1507]: time="2026-03-10T01:32:52.841120429Z" level=info msg="TearDown network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\" successfully" Mar 10 01:32:52.841355 containerd[1507]: time="2026-03-10T01:32:52.841168448Z" level=info msg="StopPodSandbox for \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\" returns successfully" Mar 10 01:32:52.843844 containerd[1507]: time="2026-03-10T01:32:52.843683730Z" level=info msg="RemovePodSandbox for \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\"" Mar 10 01:32:52.843844 containerd[1507]: time="2026-03-10T01:32:52.843738391Z" level=info msg="Forcibly stopping sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\"" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.898 [WARNING][6109] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0", GenerateName:"goldmane-cccfbd5cf-", Namespace:"calico-system", SelfLink:"", UID:"4c86706d-5198-4966-aa4f-480259ec456a", ResourceVersion:"1091", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"cccfbd5cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"c2463260a4baee1175439674d1095be1a9ea21a5088b332461387f30f0551e42", Pod:"goldmane-cccfbd5cf-sxw9t", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.57.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calid991037db6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.899 [INFO][6109] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.899 [INFO][6109] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" iface="eth0" netns="" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.899 [INFO][6109] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.899 [INFO][6109] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.944 [INFO][6116] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.944 [INFO][6116] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.945 [INFO][6116] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.955 [WARNING][6116] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.955 [INFO][6116] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" HandleID="k8s-pod-network.d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Workload="srv--eb8wl.gb1.brightbox.com-k8s-goldmane--cccfbd5cf--sxw9t-eth0" Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.957 [INFO][6116] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:52.962951 containerd[1507]: 2026-03-10 01:32:52.959 [INFO][6109] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be" Mar 10 01:32:52.962951 containerd[1507]: time="2026-03-10T01:32:52.962773506Z" level=info msg="TearDown network for sandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\" successfully" Mar 10 01:32:52.968719 containerd[1507]: time="2026-03-10T01:32:52.968685058Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:32:52.969105 containerd[1507]: time="2026-03-10T01:32:52.968895803Z" level=info msg="RemovePodSandbox \"d9c25cb3cce3fb4612f9374e933c726d75bace4d8d342c5df2589970708c49be\" returns successfully" Mar 10 01:32:52.970792 containerd[1507]: time="2026-03-10T01:32:52.970351438Z" level=info msg="StopPodSandbox for \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\"" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.047 [WARNING][6131] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"fa9659b6-52d9-4a4d-91d1-6f294264f5ee", ResourceVersion:"1112", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14", Pod:"calico-apiserver-794866df9b-h52bq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali64cc8014aee", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.048 [INFO][6131] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.048 [INFO][6131] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" iface="eth0" netns="" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.048 [INFO][6131] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.048 [INFO][6131] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.097 [INFO][6139] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.098 [INFO][6139] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.098 [INFO][6139] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.108 [WARNING][6139] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.108 [INFO][6139] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.111 [INFO][6139] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:53.119350 containerd[1507]: 2026-03-10 01:32:53.115 [INFO][6131] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.121813 containerd[1507]: time="2026-03-10T01:32:53.120340465Z" level=info msg="TearDown network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\" successfully" Mar 10 01:32:53.121813 containerd[1507]: time="2026-03-10T01:32:53.120376655Z" level=info msg="StopPodSandbox for \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\" returns successfully" Mar 10 01:32:53.121813 containerd[1507]: time="2026-03-10T01:32:53.121433463Z" level=info msg="RemovePodSandbox for \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\"" Mar 10 01:32:53.121813 containerd[1507]: time="2026-03-10T01:32:53.121471189Z" level=info msg="Forcibly stopping sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\"" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.202 [WARNING][6154] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0", GenerateName:"calico-apiserver-794866df9b-", Namespace:"calico-system", SelfLink:"", UID:"fa9659b6-52d9-4a4d-91d1-6f294264f5ee", ResourceVersion:"1112", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 31, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"794866df9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"50362763364515bd6a630352add74cba7a93c3da0b0dd000e5587a34f3998b14", Pod:"calico-apiserver-794866df9b-h52bq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.57.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-apiserver"}, InterfaceName:"cali64cc8014aee", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.202 [INFO][6154] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.203 [INFO][6154] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" iface="eth0" netns="" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.203 [INFO][6154] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.203 [INFO][6154] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.251 [INFO][6161] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.251 [INFO][6161] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.251 [INFO][6161] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.260 [WARNING][6161] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.260 [INFO][6161] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" HandleID="k8s-pod-network.50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Workload="srv--eb8wl.gb1.brightbox.com-k8s-calico--apiserver--794866df9b--h52bq-eth0" Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.262 [INFO][6161] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:53.270970 containerd[1507]: 2026-03-10 01:32:53.266 [INFO][6154] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394" Mar 10 01:32:53.270970 containerd[1507]: time="2026-03-10T01:32:53.270596217Z" level=info msg="TearDown network for sandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\" successfully" Mar 10 01:32:53.299788 containerd[1507]: time="2026-03-10T01:32:53.299706198Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:32:53.300265 containerd[1507]: time="2026-03-10T01:32:53.300123998Z" level=info msg="RemovePodSandbox \"50b80eeacac43f71fdb6ba9117dbeb12f59fa5506dd2250f2af00da241b04394\" returns successfully" Mar 10 01:32:53.301086 containerd[1507]: time="2026-03-10T01:32:53.300871631Z" level=info msg="StopPodSandbox for \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\"" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.365 [WARNING][6175] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3", ResourceVersion:"1069", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea", Pod:"coredns-66bc5c9577-2d4ds", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia236e8e9d4d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.365 [INFO][6175] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.365 [INFO][6175] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" iface="eth0" netns="" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.365 [INFO][6175] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.365 [INFO][6175] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.401 [INFO][6182] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.402 [INFO][6182] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.402 [INFO][6182] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.412 [WARNING][6182] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.412 [INFO][6182] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.416 [INFO][6182] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:53.422803 containerd[1507]: 2026-03-10 01:32:53.419 [INFO][6175] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.422803 containerd[1507]: time="2026-03-10T01:32:53.422705960Z" level=info msg="TearDown network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\" successfully" Mar 10 01:32:53.424933 containerd[1507]: time="2026-03-10T01:32:53.424092961Z" level=info msg="StopPodSandbox for \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\" returns successfully" Mar 10 01:32:53.425846 containerd[1507]: time="2026-03-10T01:32:53.425680332Z" level=info msg="RemovePodSandbox for \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\"" Mar 10 01:32:53.425846 containerd[1507]: time="2026-03-10T01:32:53.425761264Z" level=info msg="Forcibly stopping sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\"" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.483 [WARNING][6197] cni-plugin/k8s.go 616: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"ea8ee6c7-d14f-40b8-b33b-78317f32f0c3", ResourceVersion:"1069", Generation:0, CreationTimestamp:time.Date(2026, time.March, 10, 1, 30, 56, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-eb8wl.gb1.brightbox.com", ContainerID:"41d4688ff182049fed1813de2b6ddb165fa1fb8f18332121a22d622b0324e0ea", Pod:"coredns-66bc5c9577-2d4ds", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.57.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia236e8e9d4d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.485 [INFO][6197] cni-plugin/k8s.go 652: Cleaning up netns ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.485 [INFO][6197] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" iface="eth0" netns="" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.485 [INFO][6197] cni-plugin/k8s.go 659: Releasing IP address(es) ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.485 [INFO][6197] cni-plugin/utils.go 204: Calico CNI releasing IP address ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.525 [INFO][6204] ipam/ipam_plugin.go 497: Releasing address using handleID ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.525 [INFO][6204] ipam/ipam_plugin.go 438: About to acquire host-wide IPAM lock. Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.525 [INFO][6204] ipam/ipam_plugin.go 453: Acquired host-wide IPAM lock. Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.534 [WARNING][6204] ipam/ipam_plugin.go 514: Asked to release address but it doesn't exist. Ignoring ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.534 [INFO][6204] ipam/ipam_plugin.go 525: Releasing address using workloadID ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" HandleID="k8s-pod-network.42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Workload="srv--eb8wl.gb1.brightbox.com-k8s-coredns--66bc5c9577--2d4ds-eth0" Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.536 [INFO][6204] ipam/ipam_plugin.go 459: Released host-wide IPAM lock. Mar 10 01:32:53.541620 containerd[1507]: 2026-03-10 01:32:53.539 [INFO][6197] cni-plugin/k8s.go 665: Teardown processing complete. ContainerID="42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1" Mar 10 01:32:53.543371 containerd[1507]: time="2026-03-10T01:32:53.541686878Z" level=info msg="TearDown network for sandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\" successfully" Mar 10 01:32:53.546729 containerd[1507]: time="2026-03-10T01:32:53.546673597Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 10 01:32:53.546868 containerd[1507]: time="2026-03-10T01:32:53.546779354Z" level=info msg="RemovePodSandbox \"42b9b39724613b60ea9d8e632ec114941bbbf89bd20ae4d000c0814a3ef3dfb1\" returns successfully" Mar 10 01:32:55.752517 systemd[1]: Started sshd@17-10.230.50.50:22-68.220.241.50:38728.service - OpenSSH per-connection server daemon (68.220.241.50:38728). Mar 10 01:32:56.426416 sshd[6212]: Accepted publickey for core from 68.220.241.50 port 38728 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:56.430521 sshd[6212]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:56.441621 systemd-logind[1485]: New session 20 of user core. Mar 10 01:32:56.449525 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 10 01:32:57.648808 sshd[6212]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:57.661193 systemd[1]: sshd@17-10.230.50.50:22-68.220.241.50:38728.service: Deactivated successfully. Mar 10 01:32:57.661924 systemd-logind[1485]: Session 20 logged out. Waiting for processes to exit. Mar 10 01:32:57.668251 systemd[1]: session-20.scope: Deactivated successfully. Mar 10 01:32:57.671748 systemd-logind[1485]: Removed session 20. Mar 10 01:32:57.748834 systemd[1]: Started sshd@18-10.230.50.50:22-68.220.241.50:38742.service - OpenSSH per-connection server daemon (68.220.241.50:38742). Mar 10 01:32:58.357440 sshd[6227]: Accepted publickey for core from 68.220.241.50 port 38742 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:58.360261 sshd[6227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:58.373213 systemd-logind[1485]: New session 21 of user core. Mar 10 01:32:58.378411 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 10 01:32:59.195652 sshd[6227]: pam_unix(sshd:session): session closed for user core Mar 10 01:32:59.201904 systemd[1]: sshd@18-10.230.50.50:22-68.220.241.50:38742.service: Deactivated successfully. Mar 10 01:32:59.206401 systemd[1]: session-21.scope: Deactivated successfully. Mar 10 01:32:59.207796 systemd-logind[1485]: Session 21 logged out. Waiting for processes to exit. Mar 10 01:32:59.209472 systemd-logind[1485]: Removed session 21. Mar 10 01:32:59.304627 systemd[1]: Started sshd@19-10.230.50.50:22-68.220.241.50:38746.service - OpenSSH per-connection server daemon (68.220.241.50:38746). Mar 10 01:32:59.914272 sshd[6238]: Accepted publickey for core from 68.220.241.50 port 38746 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:32:59.915984 sshd[6238]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:32:59.925139 systemd-logind[1485]: New session 22 of user core. Mar 10 01:32:59.930406 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 10 01:33:01.253395 sshd[6238]: pam_unix(sshd:session): session closed for user core Mar 10 01:33:01.269324 systemd[1]: sshd@19-10.230.50.50:22-68.220.241.50:38746.service: Deactivated successfully. Mar 10 01:33:01.276736 systemd[1]: session-22.scope: Deactivated successfully. Mar 10 01:33:01.282954 systemd-logind[1485]: Session 22 logged out. Waiting for processes to exit. Mar 10 01:33:01.287751 systemd-logind[1485]: Removed session 22. Mar 10 01:33:01.364829 systemd[1]: Started sshd@20-10.230.50.50:22-68.220.241.50:38760.service - OpenSSH per-connection server daemon (68.220.241.50:38760). Mar 10 01:33:02.079351 sshd[6260]: Accepted publickey for core from 68.220.241.50 port 38760 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:33:02.085950 sshd[6260]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:33:02.100092 systemd-logind[1485]: New session 23 of user core. Mar 10 01:33:02.106462 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 10 01:33:03.245787 sshd[6260]: pam_unix(sshd:session): session closed for user core Mar 10 01:33:03.253463 systemd[1]: sshd@20-10.230.50.50:22-68.220.241.50:38760.service: Deactivated successfully. Mar 10 01:33:03.256772 systemd[1]: session-23.scope: Deactivated successfully. Mar 10 01:33:03.258180 systemd-logind[1485]: Session 23 logged out. Waiting for processes to exit. Mar 10 01:33:03.259863 systemd-logind[1485]: Removed session 23. Mar 10 01:33:03.351654 systemd[1]: Started sshd@21-10.230.50.50:22-68.220.241.50:56676.service - OpenSSH per-connection server daemon (68.220.241.50:56676). Mar 10 01:33:03.942861 sshd[6295]: Accepted publickey for core from 68.220.241.50 port 56676 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:33:03.944730 sshd[6295]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:33:03.951647 systemd-logind[1485]: New session 24 of user core. Mar 10 01:33:03.961434 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 10 01:33:04.489737 sshd[6295]: pam_unix(sshd:session): session closed for user core Mar 10 01:33:04.497080 systemd-logind[1485]: Session 24 logged out. Waiting for processes to exit. Mar 10 01:33:04.498563 systemd[1]: sshd@21-10.230.50.50:22-68.220.241.50:56676.service: Deactivated successfully. Mar 10 01:33:04.501983 systemd[1]: session-24.scope: Deactivated successfully. Mar 10 01:33:04.504818 systemd-logind[1485]: Removed session 24. Mar 10 01:33:09.605481 systemd[1]: Started sshd@22-10.230.50.50:22-68.220.241.50:56682.service - OpenSSH per-connection server daemon (68.220.241.50:56682). Mar 10 01:33:10.229566 sshd[6317]: Accepted publickey for core from 68.220.241.50 port 56682 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:33:10.233868 sshd[6317]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:33:10.245715 systemd-logind[1485]: New session 25 of user core. Mar 10 01:33:10.253462 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 10 01:33:10.853366 sshd[6317]: pam_unix(sshd:session): session closed for user core Mar 10 01:33:10.865499 systemd[1]: sshd@22-10.230.50.50:22-68.220.241.50:56682.service: Deactivated successfully. Mar 10 01:33:10.869737 systemd[1]: session-25.scope: Deactivated successfully. Mar 10 01:33:10.871273 systemd-logind[1485]: Session 25 logged out. Waiting for processes to exit. Mar 10 01:33:10.874146 systemd-logind[1485]: Removed session 25. Mar 10 01:33:15.962635 systemd[1]: Started sshd@23-10.230.50.50:22-68.220.241.50:59390.service - OpenSSH per-connection server daemon (68.220.241.50:59390). Mar 10 01:33:16.712044 sshd[6396]: Accepted publickey for core from 68.220.241.50 port 59390 ssh2: RSA SHA256:aijcv0CQPgs+ijPZDfhfY8yeUVP+ozwJgxdKg5gyU8s Mar 10 01:33:16.718482 sshd[6396]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 10 01:33:16.730448 systemd-logind[1485]: New session 26 of user core. Mar 10 01:33:16.736434 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 10 01:33:17.989415 sshd[6396]: pam_unix(sshd:session): session closed for user core Mar 10 01:33:17.999323 systemd[1]: sshd@23-10.230.50.50:22-68.220.241.50:59390.service: Deactivated successfully. Mar 10 01:33:18.008978 systemd[1]: session-26.scope: Deactivated successfully. Mar 10 01:33:18.013528 systemd-logind[1485]: Session 26 logged out. Waiting for processes to exit. Mar 10 01:33:18.017308 systemd-logind[1485]: Removed session 26.